Assignment No.9

Prototyping & Users

Updated Prototype: Image Classification on Web
http://parvezk.github.io/pre-thesis/prototype-1/index.html

Identify the community around your project and plan for feedback

Data Science
Data Scientists
Software Engineering
Software Engineer, Machine Learning
Software Engineer, Imaging Algorithm
Research Engineers
Computational Linguist
HCI Professionals
Research & Development
Research Scientists, Machine Learning & Intelligence
Visualization
Visualization Artists/Designers
Business
Business Analysts
Academia
Graduate student
PhD Student

Experts / Scholars / Researchers / Faculty

  1. Yann LeCun, Director of AI Research, Facebook, Director, NYU Center for Data Science, Silver Professor of Computer Science, Neural Science, and Electrical and Computer Engineering
  2. Enrico Bertini, Associate Professor, Computer Science and Engineering, NYU School of Engineering
  3. Anna Choromanska, Associate Professor, Computer Science and Engineering, NYU School of Engineering
  4. Kyunghyun Cho, Assistant Professor, CILVR Group, Computer Science (Courant Institute) and Center for Data Science, New York University
  5. Daniel Shiffman, Associate Arts Professor, Interactive Telecommunications Program at NYU’s Tisch School of the Arts
  6. Tega Brain, Industry Assistant Professor, NYU School of Engineering
  7. Iddo Drori, Professor, Columbia University, Dept. of Computer Science, NYU Center for Data Science
  8. Dana Karwas, Industry Assistant Professor, NYU School of Engineering
  9. Ann Borray (Messinger), Manager, ViDA Lab, NYU School of Engineering

i) List of questions for the user

  • When working on machine learning project, what type of questions concern you most? Trusting the model? Will it perform well? Will it work in deployment? predict correct output?
  • What are some deep learning model you have worked on? Can you name them?
  • Do you think linear models (such as regression models ) are easy to interpret than deep neural networks? If so, why?
  • Have you used the Matplotlib library? do you think it’s sufficient to understand, diagnose, and refine a machine learning model?
  • Have you used any other library or utility to visualize machine learning model or data?

ii) List of questions for the stakeholders

  • Do you feel the need to examine the structure of the model you work on? Do you have use any tools or method?
  • Do you think the ability to directly interact with the data and model during training process would help optimize issues such as overfitting or underfitting?
  • Do you think visualization can assist in the effective selection of network structure and other parameters?
  • How do you usually diagnose a training process that fails to converge or does not achieve an acceptable performance?
  • Would you be interested to know why some machine learning models behave the way they do and why they differ from each other?

iii) List of questions for the experts

  • With the growing complexity of AI models, do you think the critical need for understanding their inner-workings has increased?
  • Do you think visualizing the hidden activities of the artificial neural network is necessary?
  • Do you think interactive visualization would be useful in understanding and analyzing machine learning models?
  • Do you think visualizing how neural network performs its task, would give better insight for data scientists to build effective deep learning models?
  • Do you think visualizing the way neural network learns from the data would help data scientists design effective deep learning models?
  • Would you consider using a visualizer tool for understanding, debugging, and refining model?
  • Do you think integrating machine learning development with interactive visualization would improve the performance and robustness of deep learning models?
  • Would incorporating human knowledge in the training process of neural network improve the robustness of the deep learning models?

Expert Interview with Alberto Maria Chieric

Alberto Maria Chieric is an entreprenur & PhD Student at Computer Science and Engineering Department, NYU Tandon School of Engineering.

Q1: When working with the ML models, what type of questions concerns you most? whether you trust your model? Will it perform well? Will it work in deployment?
Alberto: questions that concern me: is this model the best I can do given my current data, time, and resources constraints? E.g., I cannot access or pay a GPUs EC2 instance for running a CNN or I need a result tomorrow and I know perfectly how to use a logistic regression but I would need to do some more study and trial & error for applying deep learning even if I keep reading about DL providing far better results than log reg. Also, can I deploy it already in my current architecture or will it require me re-engineering a substantial slice of production code?
Q2.What deep learning model have you used recently?
Alberto: in September I played with an RNN and a CNN in a tutorial to learn about seq2seq models for NLP and dialogue systems.
Q3: Do you feel the need to examine the structure of the deep learning model that you work on?
Alberto: yes, I do. E.g., I want to be able to come up with a strategy on how to architect the NN without going through an immense number of trials and errors. Rather, finding a rationale on why shall I chose N layers, what size each layers, what activation fn's ...etc.
Q4: Do you think visualizing how neural network performs its task, would give better insight for data scientists in building effective deep learning models?
Alberto: Yes, I believe so. E.g. if you can visualize the relative value of weights passing through neurons, and I see many times some neurons receive weights near 0, I can make a reasonable choice to eliminate a number of neurons, or to dig deeper into the neurons that get significant different than 0 weights to find out why.
Q5: Do you think visualization can assist in the effective selection of network structure, parameter selection and other optimizations?
Alberto: Intuitively I would agree because I am a visual person. However, I'm also very impatient and if I don't get a viz immediately, I tend to abandon it. E.g., I hate plots when they don't put axis labels - how am I supposed to read the plot? If it takes me more than 10-15 seconds to get it, I usually leave it. Also, it depends what exactly is visualized and my knowledge of the model. E.g., visualizing useless nodes may be helpful, also visualizing convoluted layers and what they represent in one model with N convoluted layers vs a model with K convoluted layers.
Q6: With the growing complexity of AI models, do you think the critical need for understanding their inner-workings has increased?
Alberto: Mmmh I’m not sure. I think it increased just because many use the model and regulators fear about non transparent decision making algorithms. However, in my experience I saw the companies applying models in commercial governance without worrying too much about explaining them. They even don’t worry toto much about model documentation and governance for the ones they do understand
Q7: Do you think the ability to directly interact with the data and model during training process would help optimize issues such as overfitting and underfitting?
Alberto: No sure, often one learns about under or overfitting after training is complete. It’s not quite clear to me how you can learn about over/under fitting in the middle of training. Perhaps one can check for over/under fitting with small batches and more frequently, but this already happens without the need to directly interact with the data or model during training.

User Interview with Joby Jose

Joby Jose is a Master’s Student majoring in Computer Science at NYU Tandon School of Engineering

Q1: When working on machine learning project, what type of questions concern you most? Trusting the model? Will it perform well? Will it work in deployment?
Joby: Of course, Trusting the model is top most priority.
Q2: What are some deep learning model you have worked on?
Joby: I have only done some poc on inception v3 model..planning to work on facenet as well./dd>
Q3: How do you usually diagnose a training process that fails to converge or does not achieve an acceptable performance?
Joby: I will plot the confusion matrix and change the hyper parameters and look for the improvement in confusion matrix.
Q4: Do you think visualizing how neural network performs its task, would give better insight to data scientists in building effective deep learning models?
Joby: Yes absolutely..this has infact evolved deep learning networks so much.
Q5: Would you consider using a visualizer tool for understanding, debugging, and refining model?
Joby: Well this is subjective to the problem at hand...I highly doubt that there exists a single tool that can help in visualizing all the different problems worked by DNN.
Q6: With the growing complexity of AI models, do you think the critical need for understanding their inner-workings has increased?
Joby: It's infact reduced. The fundamentals are required but to use then I guess you don't need to go deep in any network.

Assignment No.8

Prototype

Prototype: Image Classification on Web

http://parvezk.github.io/pre-thesis/prototype-1/index.html

The concept for the prototype is to build a client side artificial neural network, which can help deploy and run deep learning models on the web. We developed an image classifier that recognizes objects in the images and runs entirely in the browser, using Javascript and high-level layers API.

We use TensorFlow’s Javascript library: TensorFlow.js, a WebGL-accelerated, browser-based JavaScript library for training and deploying machine learning models on web It allows training neural networks on the browser or run pre-trained models in inference mode.

This demo uses the pre-trained MobileNet_25_224 model from Keras here.. Keras is an open source neural network library written in Python. It is designed to enable fast experimentation with deep neural networks.

It is not trained to recognize human faces. For best performance, upload images of objects like piano, coffee mugs, bottles, etc.

Assignment No.7

Outline for the Abstract Paper

  • Interpretable Machine Learning: Explore and interpret machine learning models
  • New interactive model analysis techniques and systems.
  • Method to interpret and understand deep neural networks
  • What better tools for visualizing and interpreting neural nets
  • Understand how and why a model work, how classifier works?
  • Interactive Visualization to understand, diagnose, and refine a neural network model
  • Combine recent advances in computer vision and natural language processing

Assignment No.6

Research Overview

I am profoundly excited by the idea of designing and interpreting the artificial neural networks. With the growing success of neural networks and deep learning, there is an increasing need to be able to interpret them and understand their design and inner workings of them.

What better tools would help interpret the deep neural networks and other machine learning algorithms.

Deep learning, has proved very powerful at solving problems in recent years, and it has been widely deployed for tasks like image captioning, voice recognition, and language translation. There is now hope that the same techniques will be able to diagnose deadly diseases, make trading decisions, and do countless other things to transform whole industries.

Without a clear understanding of how and why a model works, the development of these models typically relies on a time-consuming trial-and-error process. As a result, academic researchers and industrial practitioners are facing challenges that demand more transparent and explainable systems for better understanding and analyzing machine learning models, especially their inner working mechanisms.

I strongly believe it is a problem that is already relevant, and it’s going to be much more relevant in the future. Whether it’s an investment decision, a medical decision, or maybe a military decision, we don’t want to just rely on a ‘black box’ method. It’s time to act on making their decisions understandable, before the technology is even more pervasive.

I am interested in the idea of an interactive model that could help understand, diagnose, and refine a machine learning model with the help of interactive visualization.

The idea of explaining how AI technique work using an interactive model and aided with rich visualizations will not only be interesting and useful, but also discerning from the current tools and methods that involves significant mathematical notations and statistical knowledge.

I am most excited by the idea of a rich design space for interacting with the neural network. This would further help us explore their untapped potentials. It promises to be a powerful tool in enabling meaningful human oversight and in building a fair, safe, and aligned AI systems.

Assignment No.5

Dialogic Journal

Dialogic Journal on Google sheets



Link to Dialogic Journal on Google Drive


Summary of five articles

Article 1:

The Building Blocks of Interpretability

Olah, Chris, et al. The Building Blocks of Interpretability., Distill 2018, distill.pub/2018/building-blocks/.
Keywords:  Intelligibility; explanations; explainable artificial intelligence interpretable machine learning

With the growing success of neural networks and deep learning, there is an increasing need to be able to explain how neural networks make their decisions — there is a growing need to explain how they make their decisions, including building confidence about how they behave in the real-world, detecting model bias, and for scientific curiosity.

This paper discusses how we can develop tools to better interpret neural network models. This article categorizes those techniques into (1) feature visualization (2) feature attribution (3) feature grouping and shows that these can be integrated into an interactive interface as layers to let users get a sense of the internals of an image network. The article first motivates the reason to understand the network through canonical examples and visualization of concepts such as activation of neurons.

The next section talks about performing feature visualization at every spatial location across all channels. The visualizations are visually appealing, and really give a sense of how the model comes to its decision. I’m very excited to see such visualizations applied to other image datasets. Then it discuss how different concepts at different layers influence and are influenced by concepts in other layers. This is the first visualization of its kind that I’ve seen and seems like it could be extremely valuable in helping determining many details..

Beyond interfaces for analyzing model behavior, we can consider interfaces for taking action on neural networks, where the training models can learn from human feedback. Human feedback on the model’s decision-making process, facilitated by interpretability interfaces, could be a powerful solution to these problems. It might allow us to train models not just to make the right decisions, but to make them for the right reasons.

This suggests that we need more fundamental research, at the intersection of machine learning and human-computer interaction, to resolve some of these interepretability issues. Given these model behaviors are complex, author remarks that an important direction for future interpretability research will be developing techniques that achieve broader coverage of model behavior. Some other possibilities are interfaces for comparing multiple models. For instance, we might want to see how a model evolves during training.

Article 2:

Embedding Projector: Interactive Visualization and Interpretation of Embeddings

Smilkov, Daniel, et al. Embedding Projector: Interactive Visualization and Interpretation of Embeddings., ARXIV, 11/2016, arXiv:1611.05469 [stat.ML]
Keywords:  Statistics - Machine Learning, Computer Science - Human-Computer Interaction

An embedding is a map from the data to points in Euclidean space. Embeddings are everywhere in machine learning, appearing in recommender systems, Natural Language Processing, and many other applications. Researchers and developers often need to explore the properties of a specific embedding, and one way to analyze embeddings is to visualize them.

The paper presents Embedding Projector, a new visualization tool that helps users interpret machine learning models that rely on embeddings. Unlike other high-dimensional visualization systems, it is customized for the kinds of tasks faced by machine learning developers and researchers: exploring local neighborhoods for individual points, analyzing global geometry, and investigating semantically meaningful vectors in embedding space.

It can be seen that there are a number of directions for future work on visualization. For example, when developing multiple versions of a model, or inspecting how a model changes over time, it could be useful to visually compare two embeddings. The paper highlights that a second direction for future research could be make it easier for users to discover meaningful directions in the data. While the current interface makes it easy to explore various hypotheses, there may be ways for the computer to generate and test hypotheses automatically.

Article 3:

Visualizing Dataflow Graphs of Deep Learning Models in TensorFlow

Wongsuphasawat, K. (. 1. )., et al. Visualizing Dataflow Graphs of Deep Learning Models in TensorFlow., IEEE Transactions on Visualization and Computer Graphics, vol. 24 doi:10.1109/TVCG.2017.2744878.
Keywords:  Clustered Graph; Dataflow Graph; Graph Visualization; Index Terms-Neural Network

Deep learning models are becoming increasingly important in many applications. Understanding and debugging these models, however, remains a major issue. This paper describes a design study of a visualization tool that tackles one aspect of this challenge: interactive exploration of the dataflow architecture behind machine learning models.

This tool helps users understand complex machine learning architectures by visualizing their underlying dataflow graphs. The tool works by applying a series of graph transformations that enable standard layout techniques to produce a legible interactive diagram.

The design process emphasizes understanding of both users and data: It describe a task analysis for developers of deep learning models, and outline the challenges presented by model structures. The author then present a series of graph transformation to address these challenges, and demonstrate usage scenarios. They also discuss user re- actions.

Users (researchers, engineers and designers) find the visualizer useful for understanding, debugging, and sharing the structures of their models. This shows that users derived significant value from the visualizations, which is a welcome sign for visualization creators

Author concludes by saying developer reactions suggest a heartfelt desire for better ways to understand machine learning. This is an area in which data is central, but the tools have not matured, and users often feel they operate in the dark. Visualization may have an important role to play.

Article 4:

Feature Visualization

Olah, C.,Feature Visualization  [link], Distill, 2017 DOI: 10.23915/distill.00007
Deep Learning, Neural Network, GoogLeNet Visualizations, Interactive Visualizations

Neural networks are algorithms inspired by the biological brain. Deep learning is a set of techniques for learning in neural networks. They have made a huge impact on a wide variety of applications recently. They provide some of best solutions to many problem in image recognition, voice recognition and autonomous driving.

While deep neural networks learn efficient and powerful representations, they are often considered a ‘black-box’. With their growing influence in our lives, there is a critical need to better understand their decisions and to gain insight into how these models operate. This paper attempts to interpret a neural network model for an image classification task.

Feature visualization is one of the important techniques to understand what neural networks have learned from image dataset. This paper focuses on optimization methods and discusses major issues and explore common approaches to solving them.

The author presents a comprehensive overview on feature visualization, mainly focusing on optimization method. It starts with explaining why use optimization method to visualize feature in neural networks compared to finding examples from the dataset. Then discuss how to achieve diversity with optimization. Then further discuss the interaction between neurons, which can explore the combinations of neurons working together to represent images in neural networks. Finally, discuss how to improve the optimization process better by adding different regularizations methods.

The author concludes by saying feature visualization stands out as one of the most promising and developed research directions. By itself, feature visualization might never give a completely satisfactory understanding. But as one of the fundamental building blocks when, combined with additional tools, will empower humans to understand these systems.

Article 5:

Show and Tell Image Captioning Challenge

Vinyals, O., et al.Show and Tell: Lessons Learned from the 2015 MSCOCO Image Captioning Challenge., IEEE Transactions on Pattern Analysis and Machine Intelligence, no. 4, 2017.
Image captioning, recurrent neural network, sequence-to-sequence, language model.

This paper presents a machine learning model, which is trained to describe the content of an image. Until recently, it was a fundamental problem in the artificial intelligence area that connects two emerging fields: computer vision and natural language processing.

The work is inspired by the recent advances in machine translation such as Google translate application. For many years, machine translation was achieved by a series of separate tasks, but recent work has shown that translation can be done in a much simpler way and still reach greater performance.

The paper proposes a a neural and probabilistic framework to generate descriptions from images. These models make use of an architecture called called Recurrent Neural Network (RNN) that combines recent advances in computer vision and machine translation and that can be used to generate natural sentences describing an image. The model is trained to maximize the likelihood of the target description given the training image.

They present an end-to-end system for the problem. It is a neural net which is fully trainable using stochastic gradient descent. then, their model combines sub-networks for vision and language models. They can be pre-trained on larger corpora and thus can take advantage of additional data.

This paper was the result of a competition for an image captioning challenge, organized by Microsoft in 2015, where the team contested and won and placed equal first with Microsoft Research. The authors discusses the resulting performance in the competition and experiments on several datasets.

Assignment Week 4

Step 1 - Big Problem

With the growing complexity of artificial neural networks, the critical need for understanding their inner-workings has increased. It is being used to guide all sorts of key decisions in medicine, finance, manufacturing and beyond, including determining who gets healthcare, who’s approved for a loan and who gets hired for a job.

Step 1A - Big Problem with Data

  • Machine Learning is a subset of artificial intelligence. It allows the machines to learn and make predictions based on its experience (data).
  • Machine learning has been successfully applied to a wide variety of fields ranging from information retrieval, data mining, and speech recognition, to computer graphics, visualization, and human–computer interaction. However, most users often treat a machine learning model as a black box because of its incomprehensible functions and unclear working mechanism.
  • The AI technology, known as deep learning, has proved very powerful at solving problems in recent years, and it has been widely deployed for tasks like image captioning, voice recognition, and language translation. There is now hope that the same techniques will be able to diagnose deadly diseases, make trading decisions, and do countless other things to transform whole industries.
  • Without a clear understanding of how and why a model works, the development of these models typically relies on a time-consuming trial-and-error process. As a result, academic researchers and industrial practitioners are facing challenges that demand more transparent and explainable systems for better understanding and analyzing machine learning models, especially their inner working mechanisms.
  • “It is a problem that is already relevant, and it’s going to be much more relevant in the future,” says Tommi Jaakkola, a professor at MIT who works on applications of machine learning. “Whether it’s an investment decision, a medical decision, or maybe a military decision, you don’t want to just rely on a ‘black box’ method.”. It’s time to act on making their decisions understandable, before the technology is even more pervasive.

Step 2A - Stakeholder map

Step 2B - Stakeholder list

  • Data Scientists
  • AI Researchers
  • Machine Learning Engineers
  • PhD & Post-doctoral students
  • Research & Development centers
  • Stock Traders
  • Economic Analysts
  • University Lab & centres

Step 3A Personas

Data Scientists

Data scientists and engineers, who are facing challenges that demand more transparent and explainable systems for better understanding and analyzing machine learning models in their projects, especially their inner working mechanisms.

An interactive visualization of these models is needed to effectively solve some of the real world challenges. An interactive model could help data scientist understand, diagnosing, and refining a machine learning model with the help of interactive visualization.

Step 3B - User List

  • Chetan N.G, System Engineers, NVIDIA
  • Enrico Bertini, Assistant Professor, Department of Computer Science and Engineerin, NYU Tandon School of Engineering
  • Xianbo Xi, PhD Student, Tandon School of Engineering

Assignment No.3

Step 1: Determining Area & Topic

Area: Machine Learning

Topic: Neural Networks

Brainstorming, freewriting, and mind mapping

Brainstorm

Freewrite

Mind map

Step 2

Question - Response - Question



Augmenting Human Intellect

  • How can we use artificial intelligence to augment human intellect?
  • Can machine learning help augment human creativity?
  • How generative interfaces can be used to explore and visualize machine learning models?
  • Will these tools help humans think and create in new ways?
  • Can they be used to generate ideas, which are truly original?
  • Can such systems be used to develop fundamental new interface primitives?
  • How will those new primitives change and expand the way humans think?


Explainable AI

  • How to explain how AI system works?
  • What are the dark secrets of AI systems?
  • With their growing complexity and pervasiveness, how do we interpret the inner workings of AI models?
  • Can design thinking help solve the issue of explainability?
  • Will research at the intersection of HCI and AI help resolve these issues?
  • Can visualization techniques help fill such critical need?
  • Can visual narratives help bring new insight into the complexity of AI system?


Why Neural Network is a black box?

  • How neural network learns on its own from the data?
  • What does the network see?
  • Can we see what the network detects?
  • How it assembles various individual pieces as it moves forward?
  • Can we detect model bias?
  • How does it arrive at the final decision?
  • what are the relationships like between different neurons?

Wheel and Spoke

Step 3: List of questions (Factual, Technical, Conceptual)

Factual

  • How can we enable human potential using emerging technologies?
  • How can we augment human intellect with artificial intelligence
  • What does the future hold for new field emerging today out of a synthesis of AI and IA?
  • What type of research would be useful at the intersection of machine learning and human-computer interaction

Technical

  • How to build visual interface to interpret machine learning algorithms?
  • How do we interpret high dimensional data in a visual medium?

Conceptual

  • How do we create a space to interact with neural networks?
  • Can we trust deep learning techniques?
  • How do we use interpretability methods as powerful tool in enabling meaningful human oversight and in building fair, safe, and aligned AI systems

Accountability Partner

Alex Nathanson and Jason Charles are my accountability partners. We will be meeting briefly before pre-thesis every Tuesday.

Assignment No.2

Journal Articles

Review Articles

Article Response

The five articles I read were:

  • Throwing shapes
  • Why the Focus on Human Intelligence?
  • Mapping’s Intelligent Agents
  • Plumbing the Depths of Neural Nets
  • Interactive Visualization and Interpretation of Embeddings

Responses

Visual Experiments

Branches of Aritifical Intelligence

Network

Two layer Neural Network

Article 1:

Throwing shapes

The author examines that in order to discover how the brain works, we need to look to higher dimensions. The research states as our brains think, learn and remember, they create elaborate but ephemeral structures in mathematical dimensions. These structures which appear and disappear, could help us understand how the brain creates our thoughts and feelings.

  • What does this do for me?

    The research finding intrigued me where they have been using algebraic topology, a field of mathematics used to characterize higher- dimensional shapes, to explore the workings of the brain.

  • What is interesting about this article?

    The blue brain project launched by the researchers aimed to simulate the entire human brain inside a computer. Not only have did they managed to do all their research in computer but they have also managed to replicate some if it’s finding through actually biological research.

  • What questions is this making me ask?

    What are those higher dimension structures? What if the geometric structures they represent exist in mathematical dimensions higher than we can visualize?

Article 2:

Why the Focus on Human Intelligence?


The article discusses the idea to reverse engineer the brain mechanisms that underlie human visual intelligence.

  • What does this do for me?

    The idea that a better understanding of the brain could lead to better computer algorithms and AI deepened my interest to learn more about the brain mechanism.

  • What is interesting about this article?

    The Beautiful Brain, a 3-D interactive visualization based on the drawings of Santiago Ramón y Cajal interested me with the visual aspect of the human intelligence.

  • What questions is this making me ask?

Article 3:

Mapping’s Intelligent Agents


The article examines autonomous vehicles, sensing technologies, indigenous cartography, and non-human spatial sensibilities.

  • What does this do for me?

    The capacities and limitations of machine-mapping and autonomous spatial technologies needs attention.

  • What is interesting about this article?

    Mapping made for and by machines is a big business opportunity at present. Yet mapping’s artificial intelligences also have the potential to transform myriad of design and research areas, to influence policy-making and governance, to support environmental preservation and public health.

  • What questions is this making me ask?

    How machines conceptualize and operationalize space? How do they render our world measurable, navigable, usable, conservable?

Article 4:

Plumbing the Depths of Neural Nets


Tomaso Poggio, the director of Center for Brains, Minds, and Machines (CBMM) at MIT’s McGovern Institute for Brain Research discuss the implications for human intelligence.

  • What does this do for me?

    The artificial neural networks, is modeled loosely after the human brain, They can train themselves to recognize complex patterns, and provide solutions to many problems in image recognition, speech recognition, and natural language processing.

  • What is interesting about this article?

    The author explains that problem of intelligence is not only the engineering problem of building an intelligent machine, but the scientific problem of what intelligence is, how our brain works and how it creates the mind.

  • What questions is this making me ask?

    What are the major technology trends driving Deep Learning? How to build, train and apply fully connected deep neural networks.

Article 5:

Interactive visualization of high-dimensional datasets


Embedding Projector, a system for interactive visualization and analysis of high-dimensional datasets used in machine learning projects. Researchers and developers often need to explore the properties of a specific embedding, and one way to analyze embedding is to visualize them.

  • What does this do for me?

    A new tool that could help users interpret machine learning models visually could help reduce common problems in machine leaning such as overfitting and underfitting that would eventually lead to better accuracy.

  • What is interesting about this article?

    How the embedding projector helps reduce the arbitrary high-dimensional data into a simple two- or three-dimensional view is interesting.

  • What questions is this making me ask?

    What are the major technology trends driving Deep Learning? How does it apply dimensionality reduction techniques such as t-SNE and PCA for an arbitrary dataset.

Inspiration Cards

Areas of Interest

Inspiration cards

Personal Interest Cards

Neural Networks

Power of Algorithms

Prediction Machines

Dark Matters

API for Non-Profits

Questionniare

Initial Review Investigations

How neural network works?

  • How are neural networks formed?
  • What are the types of neural network?
  • How neurons firing in our brain create complex web of connections?
  • What are artifical neural networks
  • How does human mind conjures our sense of self?

What are some powerful
algorithms?

Algorithms are already all around us. It is what drives Google’s search engine, powers Netflix to figure out what we want to watch next, Amazon Alexa’s voice assistants, Tinder’s match-making, autonomous vehicles, high-speed trading and an ever-growing number of services and technologies.

  • what is the ultimate algorithm with the potential to remake our world
  • How to identify and highlight algorithmic bias?
  • How to drive diversity and inclusiveness in today’s predictive decision-making tools?
  • How can we bring more awareness about the negative impact of these automated systems?

What does AI mean for
you and me?

Artificial Intelligence will affect everyone, everywhere. The past few years have seen rapid advances in AI, with new technologies achieving dramatic improvements in technical performance.

Where does this ultimately lead? Software that thinks and does. Software with cognitive ability that predict people’s behavior and drive autonomous vehicles etc.

  • Would prediction tools help improve productivity in our lives?
  • Will it help reduce uncertainties in our day-to-day decision-making?
  • Would Artificial Intelligence code a better world? A better future?
  • Would it help to build a better, smarter and more connected world?
  • What are the most important trend in the world of AI?
  • Can design thinking open up new applications for AI?
  • How might we ensure inclusion so that everyone can benefit from breakthroughs in AI?

Is Dark Matter Real?

  • Is dark matter real or just a hypothetical form?
  • Why do physicists beleive dark matter do not interact with light particles
  • Is mathematical proof sufficient to satisfy it's existence
  • How long has dark matter been around?
  • Is it related to Dark Energy?

How to create APIs for Non-Profit

APIs are all about sharing. It's an interface that lets one software program “talk” to another, exchanging data behind the scenes. APIs power much of what we do online. Startups and established companies alike are leveraging the technology to become a bigger part of our lives.

  • How to to use API ecosystem to elevate non-profit works?
  • How to harness the power of APIs for greater good or social causes?
  • How can we use APIs to help non-profit organizations to wider their impact?
  • How to leverage APIs from multiple open source platforms to build innovative social services?
  • How to use APIs to collect charitable donations