TensorFlow is an open source framework, by Google for creating deep learning models. Deep Learning is one of several categories of machine learning (ML) models that use multi-layer neural networks. TensorFlow is cross-platform. It runs on nearly everything: GPUs and CPUs—including mobile and embedded platforms—and even tensor processing units (TPUs), which are specialized hardware to do tensor math on The TensorFlow library allows users to perform functions by creating a computational graph.
TensorFlow provides a variety of different toolkits that allow you to construct models at your preferred level of abstraction. You can use lower-level APIs to build models by defining a series of mathematical operations
Tensor Flow is a free and open source framework. It was originally developed by the Google Brain team. It is mainly used for numerical computation using data flow graphs.
TFT stands for TensorFlow Transform. It is an open source library for TensorFlow that allows users to define pre-processing pipelines and run these using large scale data processing frameworks, while also exporting the pipeline in a way that can be run as part of a TensorFlow graph. Users define a pipeline by composing modular Python functions, which tf.Transform then executes with Apache Beam, a framework for large-scale, efficient, distributed data processing.
Eager execution provides an imperative interface to TensorFlow (It is similar to NumPy). When you enable eager execution, TensorFlow operations execute immediately; you do not execute a pre-constructed graph with Session.run().Eager mode is moving out of contrib, using eager execution you can run your code without a session. Eager execution is a flexible machine learning platform for research and experimentation, providing: An intuitive interface:Structure your code naturally and use Python data structures. Quickly iterate on small models and small data. Easier debugging:Call ops directly to inspect running models and test changes. Use standard Python debugging tools for immediate error reporting. Natural control flow:Use Python control flow instead of graph control flow, simplifying the specification of dynamic models. Eager execution runs by default on CPU, to use GPU include below code: with tf.device(‘/gpu:0’)  Eager execution doesn’t create Tensor Graph, to build graph just remove the tf.enable_eager_execution().
Eager Execution is compatible with native Python debugging tools Error logging is immediate Native Python control flow i.e loops and recursions Eager execution simplifies your code Back propagation is built in to eager execution
It is possible to replace them with other values that run parallel to them. Dropna and isnull are the two methods that are useful in this matter. In some special cases, it is even possible to replace them with desired values and have an error-free outcome.
Tensorflow is one of many machine learning libraries (other examples include CNTK and Theano). Machine learning is a field of computer science that gives computers the ability to learn without being explicitly programmed.[1] For example, AlphaGo Zero‘s AI taught itself to play Go and outperformed its predecessor, AlphaGo, which had defeated the world champion in Go. Machine learning is useful for tasks where explicit algorithms don’t yield good results, such as user screening, sorting high context data and clustering for predictions and profiling.Practical examples include detecting fraud or data breaches, email filtering, optical character recognition (OCR), and ranking.
It is basically a sub-algorithm of a sub-module that defines the conditional probabilities of different components. The final results can be integrated with other possible outcomes to predict the final outcomes. It can also overcome a lot of problems which are related to the unstructured data.
Type I error is a false positive value. On the other side, Type II error is a false negative value. Type I error generally represent that something has happened when actually it doesn’t while Type II error is to representing the machine that nothing is wrong when actually something is not good.
This theorem defines the probability of any event in machine learning. It represents a fixed value which is actually a mathematical calculation. This value is generally obtained by dividing the true positive rate divided by the false positive rate. In machine learning, some of the very complex problems and challenges can easily be solved and eliminated with the help of this theorem. Most of the time results provided by it are highly accurate and can easily be trusted.
K-means clustering is basically an unsupervised clustering algorithm. It is capable to tolerate some minor errors. On the other side, the KNN is structured clustering algorithm. For reliable operations, it should be accurate and reliable. The mechanism for both seems very similar at the first glance but users need to label the data in the KNN which is not required in the k-means clustering.
Tensorflow Loaders are used for adding algorithms and data backends one of which is tensorflow itself. For example, a loader can be implemented to load, access and unload a new type of servable machine learning model
ROC or region of convergence used to reflect data rates which classify as true positive and false positive. Represented in the form of graphs, it can use as a proximity to swap operations related to different algorithms.
Tensorflow can also use with containerization tools such as docker, for instance, it could use to deploy a sentiment analysis model which uses character level ConvNet networks for text classification.
Sequence-to-Sequence (Seq2Seq) models use recurrent neural networks as a building block by feeding lots of sentence pairs during model training so that we can generate one sentence from another sentence. These sentence pairs can be anything. For example, when it is words from two different languages, the model can be used for translations. When it is a pair of conversational messages, the model can be used for chat bots.
Keras is a high-level library that’s built on top of Theano or TensorFlow. It provides a scikit-learn type API for building Neural Networks. It is written in Python programming language.Developers can use Keras to quickly build neural networks without worrying about the mathematical aspects of tensor algebra, numerical techniques, and optimisation methods. It’s used for fast prototyping, state-of-the-art research, and production. While TensorFlow supports Keras today, with 2.0, we are integrating Keras more tightly into the rest of the TensorFlow platform.
A max pooling layer performs down-sampling by dividing the input into rectangular pooling regions, and computing the maximum of each region. having higher levels of abstraction.
TFLearn: This API shouldn’t be seen as TF Learn, which is TensorFlow’s tf.contrib.learn. It is a separate Python package.TensorLayer: It comes as a separate package and is different from what TensorFlow’s layers API has in its bag. Pretty Tensor: It is actually a Google project which offers a fluent interface with chaining. Sonnet: It is a project of Google’s DeepMind which features a modular approach
TensorFlow operator     Description tf.add                  x+y tf.subtract          x-y tf.multiply          x*y tf.div                    x/y tf.mod                 x % y tf.abs                   |x| tf.negative         -x tf.sign                  sign(x) tf.square            x*x tf.round              round(x) tf.sqrt                  sqrt(x) tf.pow                 x^y tf.exp                   e^x tf.log                    log(x) tf.maximum      max(x, y) tf.minimum       min(x, y) tf.cos                    cos(x) tf.sin                     sin(x)
Sequence utilities include methods such as: argmin function,:which returns the index with min value across the axes of the input tensor, argmax function: which returns the index with max value across the axes of the input tensor, setdiff: which computes the difference between two lists of numbers or strings, where function:which will return elements either from two passed elements x or y, which depends on the passed condition, or unique function,:which will return unique elements in a 1-D tensor.
Reduction is an operation that removes one or more dimensions from a tensor by performing certain operations across those dimensions.
An Embedding in TensorFlow defines as the mapping like the word to vector (word2vec) of real numbers. A TensorFlow embedding example below where a list of colors represents as vectors: Black: (0.01359, 0.00075997, 0.24608, …, -0.2524, 1.0048, 0.06259) Blues: (0.01396, 0.11887, -0.48963, …, 0.033483, -0.10007, 0.1158) Yellow: (-0.24776, -0.12359, 0.20986, …, 0.079717, 0.23865, -0.014213) Oranges: (-0.35609, 0.21854, 0.080944, …, -0.35413, 0.38511, -0.070976) The dimensions in these kinds of vectors usually don’t have any meaning, but the pattern of the matrix, as well as the location and distance between the vectors, contains some significant information that can be taken an advantage of.
TensorFlow Mobile is used for a mobile platform such as iOS and Android. This is for those developers who have a successful TensorFlow model and want to integrate their model into a mobile environment. This is also for those who are not able to use TensorFlow Lite. Basic challenges one can find in integrating their desktop environment model into the mobile environment are:
Data Formats is one of the ways for TensorFlow Performance Optimizations. As the name suggests, the structures of the input tensors that passes to the operations. Below are parameters of a 4D tensor: The nomenclature of these data formats in broadly split into two: The default is the latter whereas the firmer is the optimal choice while working on NVIDIA GPUs. It is good to build models that are compatible with both the formats so as to simplify the training on GPUs.
There are two configurations below for optimizing CPU performance. intra_op_parallelism: parallelization of nodes is achieved using multi-threads, that’ll schedule the individual pieces. inter_op_parallelism: The nodes that are ready are scheduled in this operation. tf.ConfigProto, is used to set these configurations by passing to in the config attribute of tf.Session. For both parallelism configurations, is initialised as zero, they will revert to the number of logical CPU cores. Equating the number of physical cores to the number of threads rather than using logical cores is another good way of optimization.
MNIST dataset in TensorFlow, containing information of handwritten digits spiltted into three parts: Training Data (mnist.train) – 55000 datapoints Validation Data (mnist.validate) – 5000 datapoints Test Data (mnist.test) – 10000 datapoints
Tenosrflow has its own debugger called the tfdbg TensorFlow Debugging, which lets you observe the intrinsic working and the state of the running graphs. These are difficult to debug with any other general debuggers like pdb in python.
What is TensorFlow?
What is TFT?
Explain Eager Execution?
What are the Features of Eager Execution?
What if a file is corrupted or missing in a dataset?
How does TensorFlow fit into AI and Machine learning?
What is the application of Naïve Bayes Naïve in Machine Learning?
What difference do you find in type1 and type 2 errors?
How useful and reliable Bayes’ theorem is according to you in the Machine Learning context?
How K-means clustering is different from KNN?
What are TensorFlow loaders?
What is ROC curve and its working?
Can TensorFlow be deployed in container software?
What is Sequence-to-Sequence model?
What is Keras?
What is Max pooling?
What are the APIs outside TensorFlow project?
What are the TensorFlow operations?
What are the Sequence Utilities methods?
What is Reduction?
What are the cons of TensorFlow?
Can you explain Embedding in TensorFlow?
What is TensorFlow Mobile?
Can you explain Data Formats in TensorFlow?
Explain TensorFlow Optimizing for CPU?
What is MNIST Dataset in TensorFlow?
What is the tfdbg TensorFlow Debugging?
What are the Common TensorFlow Operations?