This course aims to give you an easy to understand guide to the complexities of Google's TensorFlow 2 framework in a way that is easy to understand.We'll focus on understanding the latest updates to TensorFlow and leveraging the Keras API (TensorFlow 2.0's official API) to quickly and easily build models. In this course we will build models to forecast future price homes, classify medical images, predict future sales data, generate complete new text artificially and much more! This course is designed to balance theory and practical implementation, with complete jupyter notebook guides of code and easy to reference slides and notes. We also have plenty of exercises to test your new skills along the way! Learn to use TensorFlow 2.0 for Deep Learning Leverage the Keras API to quickly build models that run on Tensorflow 2 Perform Image Classification with Convolutional Neural Networks Use Deep Learning for medical imaging Forecast Time Series data with Recurrent Neural Networks Use Generative Adversarial Networks (GANs) to generate images Use deep learning for style transfer Generate text with RNNs and Natural Language Processing Serve Tensorflow Models through an API Use GPUs for accelerated deep learning This course will guide you through how to use Google's latest TensorFlow 2 framework to create artificial neural networks for deep learning!
This article aims to provide the basic knowledge of how to recognize snacks by using Python and SimpleCV. Readers will gain practical programming knowledge via experimentation with the Python scripts included in the Snack Classifier open source project. To illustrate with a snacks recognition app, the Snack Watcher watches any snacks present on the snack table. For Snack Watcher to determine if there was an interesting event, it needs to process the image into a set of image "Blobs". For each "Blob", Snack Watcher compares the "Blob" with it's previous state to determine if the "Blob" was added, removed or stationary.
We present an interactive system enabling users to manipulate images to explore the robustness and sensitivity of deep learning image classifiers. Using modern web technologies to run in-browser inference, users can remove image features using inpainting algorithms and obtain new classifications in real time, which allows them to ask a variety of "what if" questions by experimentally modifying images and seeing how the model reacts. Our system allows users to compare and contrast what image regions humans and machine learning models use for classification, revealing a wide range of surprising results ranging from spectacular failures (e.g., a "water bottle" image becomes a "concert" when removing a person) to impressive resilience (e.g., a "baseball player" image remains correctly classified even without a glove or base). We demonstrate our system at The 2018 Conference on Computer Vision and Pattern Recognition (CVPR) for the audience to try it live. Our system is open-sourced at https://github.com/poloclub/interactive-classification. A video demo is available at https://youtu.be/llub5GcOF6w.