Commit a9663527 authored by Peter Eastman's avatar Peter Eastman
Browse files

Merged changes from master branch

parents e87b9dd7 8233320d
Loading
Loading
Loading
Loading
+2 −2
Original line number Diff line number Diff line
@@ -21,9 +21,9 @@ ENV PATH /miniconda/bin:$PATH
# TODO: Get rid of this when there is a stable release of deepchem.
RUN git clone https://github.com/deepchem/deepchem.git && \
    cd deepchem && \
    git checkout 415aebadff54175b7ba108964723c8f69438af94 && \
    git checkout tags/1.0.0 && \
    bash scripts/install_deepchem_conda.sh root && \
    pip install tensorflow-gpu==0.12.1 && \
    pip install tensorflow-gpu==1.0.1 && \
    python setup.py develop

# Clean up
+6 −2
Original line number Diff line number Diff line
@@ -44,7 +44,7 @@ You can install deepchem in a new conda environment using the conda commands in

```bash
bash scripts/install_deepchem_conda.sh deepchem
pip install tensorflow-gpu==0.12.1                      # If you want GPU support
pip install tensorflow-gpu==1.0.1                      # If you want GPU support
git clone https://github.com/deepchem/deepchem.git      # Clone deepchem source code from GitHub
cd deepchem
python setup.py install                                 # Manual install
@@ -95,7 +95,7 @@ via this installation procedure.
    contact your local sysadmin to work out a custom installation. If your
    version of Linux is recent, then the following command will work:
    ```
    pip install tensorflow-gpu==0.12.1
    pip install tensorflow-gpu==1.0.1
    ```

9. `deepchem`: Clone the `deepchem` github repo:
@@ -484,3 +484,7 @@ Approaches](http://pubs.acs.org/doi/abs/10.1021/acs.jcim.6b00290)

## About Us
DeepChem is a package by the [Pande group](https://pande.stanford.edu/) at Stanford. DeepChem was originally created by [Bharath Ramsundar](http://rbharath.github.io/), and has grown through the contributions of a number of undergraduate, graduate, and postdoctoral researchers working with the Pande lab.


## Version
1.0.1
+2 −0
Original line number Diff line number Diff line
@@ -9,6 +9,8 @@ from deepchem.models.models import Model
from deepchem.models.sklearn_models import SklearnModel
from deepchem.models.tf_new_models.multitask_classifier import MultitaskGraphClassifier
from deepchem.models.tf_new_models.multitask_regressor import MultitaskGraphRegressor
from deepchem.models.tf_new_models.DTNN_regressor import DTNNGraphRegressor

from deepchem.models.tf_new_models.support_classifier import SupportGraphClassifier
from deepchem.models.multitask import SingletaskToMultitask
from deepchem.models.sequential import Sequential
+4 −3
Original line number Diff line number Diff line
@@ -631,8 +631,9 @@ class TensorflowClassifier(TensorflowGraphModel):
      A tensor with shape batch_size containing the weighted cost for each
      example.
    """
    return tf.mul(
        tf.nn.softmax_cross_entropy_with_logits(logits, labels), weights)
    return tf.multiply(
        tf.nn.softmax_cross_entropy_with_logits(logits=logits, labels=labels),
        weights)

  def add_label_placeholders(self, graph, name_scopes):
    """Add Placeholders for labels for each task.
@@ -785,7 +786,7 @@ class TensorflowRegressor(TensorflowGraphModel):
      A tensor with shape batch_size containing the weighted cost for each
      example.
    """
    return tf.mul(0.5 * tf.square(output - labels), weights)
    return tf.multiply(0.5 * tf.square(output - labels), weights)

  def add_label_placeholders(self, graph, name_scopes):
    """Add Placeholders for labels for each task.
+4 −3
Original line number Diff line number Diff line
@@ -35,7 +35,7 @@ def weight_decay(penalty_type, penalty):
    else:
      raise NotImplementedError('Unsupported penalty_type %s' % penalty_type)
    cost *= penalty
    tf.scalar_summary('Weight Decay Cost', cost)
    tf.summary.scalar('Weight Decay Cost', cost)
  return cost


@@ -144,8 +144,9 @@ class TensorflowLogisticRegression(TensorflowGraphModel):
      return loss

  def cost(self, logits, labels, weights):
    return tf.mul(
        tf.nn.sigmoid_cross_entropy_with_logits(logits, labels), weights)
    return tf.multiply(
        tf.nn.sigmoid_cross_entropy_with_logits(logits=logits, labels=labels),
        weights)

  def add_output_ops(self, graph, output):
    # adding output nodes of sigmoid function
Loading