Commit c9687ca5 authored by leswing's avatar leswing
Browse files

Formatting

parent 66cef953
Loading
Loading
Loading
Loading
+12 −12
Original line number Diff line number Diff line
@@ -183,6 +183,7 @@ class Input(Layer):
  def __call__(self, *parents):
    self.out_tensor = tf.placeholder(tf.float32, shape=self.t_shape)


class LossLayer(Layer):

  def __init__(self, **kwargs):
@@ -247,10 +248,8 @@ class ReduceMean(Layer):


class Conv2d(Layer):
  def __init__(self,
               num_outputs,
               kernel_size=5,
               **kwargs):

  def __init__(self, num_outputs, kernel_size=5, **kwargs):
    self.num_outputs = num_outputs
    self.kernel_size = kernel_size
    super().__init__(**kwargs)
@@ -268,6 +267,7 @@ class Conv2d(Layer):


class MaxPool(Layer):

  def __init__(self,
               ksize=[1, 2, 2, 1],
               strides=[1, 2, 2, 1],
+13 −15
Original line number Diff line number Diff line
@@ -16,6 +16,7 @@ from deepchem.trans import undo_transforms


class TensorGraph(Model):

  def __init__(self,
               tensorboard=False,
               tensorboard_log_frequency=100,
@@ -223,10 +224,7 @@ class TensorGraph(Model):
        y_pred = np.reshape(y_pred, (n_samples, n_tasks))
        return y_pred

  def predict_proba(self,
                    dataset,
                    transformers=[],
                    batch_size=None):
  def predict_proba(self, dataset, transformers=[], batch_size=None):
    """
    TODO: Do transformers even make sense here?

+0 −1
Original line number Diff line number Diff line
@@ -41,7 +41,6 @@ class TestTensorGraph(unittest.TestCase):
    g.add_layer(loss, parents=[dense, label_out])
    g.set_loss(loss)


    g.fit(dataset, nb_epoch=100)
    g.save()
    g1 = TensorGraph.load_from_dir('/tmp/tmpss5_ki5_')