Thank you for your interest to our library ! We are thrilled to have such a motivated community.
- If you're new with TFDS, the easiest way to get started is to implement one of our requested dataset, focusing on the most requested once. Follow our guide for instructions.
- Issues, feature requests, bugs,... have a much bigger impact than adding new datasets, as they benefit the entire TFDS community. See the potential contribution list. Starts with the ones labeled with contribution-welcome which are small self-contained easy issues to get started with.
- Don't hesitate to take over bugs which are already assigned, but haven't been updated in a while.
- No need to get the issue assigned to you. Simply comment on the issue when you're starting working on it :)
- Don't hesitate to ask for help if you're interested by an issue but don't know how to get started. And please send draft PR if you want early feedback.
- To avoid unnecessary duplication of work, check the list of pending Pull Requests, and comment on issues you're working on.
Cloning the repo
To get started, clone or download the Tensorflow Datasets repository and install the repo locally.
git clone https://github.com/tensorflow/datasets.git cd datasets/
Install the development dependencies:
pip install -e . # Install minimal deps to use tensorflow_datasets pip install -e ".[dev]" # Install all deps required for testing and development
Note there is also a
pip install -e ".[tests-all]" to install all
Visual Studio Code
- If you are encountering some TensorFlow warning message, try this fix.
- If discovery fail due to missing import which should have been installed,
please send a PR to update the
Sign the CLA
Ensure you have signed the Contributor License Agreement (CLA), so we can accept your contributions (only required once for all Google projects).
Follow best practices
- Readability is important. Code should follow best programming practices (avoid duplication, factorise into small self-contained functions, explicit variables names,...)
- Simpler is better (e.g. implementation split into multiple smaller self-contained PRs which is easier to review).
- Add tests when required, existing tests should be passing.
- Add typing annotations
Check your style guide
- Line length: 80
- 2 spaces indentation instead of 4.
- Single quote
pip install pylint --upgrade pylint tensorflow_datasets/core/some_file.py
You can try
yapf to auto-format a file, but the tool is not perfect, so you'll
likelly have to manually apply fixes afterward.
yapf should have been installed with
pip install -e
".[dev]" but can also be manually installed with
pip install. If you're using
VS Code, those tools should be integrated in the UI.
Docstrings and typing annotations
Classes and functions should be documented with docstrings and typing annotation. Docstrings should follow the Google style. For example:
def function(x: List[T]) -> T: """One line doc should end by a dot. * Use `backticks` for code and tripple backticks for multi-line. * Use full API name (`tfds.core.DatasetBuilder` instead of `DatasetBuilder`) * Use `Args:`, `Returns:`, `Yields:`, `Attributes:`, `Raises:` Args: x: description Returns: y: description """
Add and run unittests
Make sure new features are tested with unit-tests. You can run tests through the VS Code interface, or command line. For instance:
pytest -vv tensorflow_datasets/core/
unittest: Historically, we have been using
unittest module to
write tests. New tests should preferably use
pytest which is more simple,
flexible, modern and used by most famous libraries (numpy, pandas, sklearn,
matplotlib, scipy, six,...). You can read the
if you're not familiar with pytest.
Tests for DatasetBuilders are special and are documented in the guide to add a dataset.
Send the PR for reviews!
Congrats! See GitHub Help for more information on using pull requests.