Author Archives: pgfeldman

Phil 4.16.18

9:00 – ASRC MKT

  • Finished up and submitted the CI 2018 and also put up on ArXive. Probably 90 minutes total?
  • SASO deadlines got extended:
    • Abstract submission (extended)  April 23, 2018
    • Submission (extended) April 30, 2018
  • Some diversity injection: Report for America Supports Journalism Where Cutbacks Hit Hard
    • Report for America, a nonprofit organization modeled after AmeriCorps, aims to install 1,000 journalists in understaffed newsrooms by 2022. Now in its pilot stage, the initiative has placed three reporters in Appalachia. It has chosen nine more, from 740 applicants, to be deployed across the country in June.
  • An information-theoretic, all-scales approach to comparing networks
    • As network research becomes more sophisticated, it is more common than ever for researchers to find themselves not studying a single network but needing to analyze sets of networks. An important task when working with sets of networks is network comparison, developing a similarity or distance measure between networks so that meaningful comparisons can be drawn. The best means to accomplish this task remains an open area of research. Here we introduce a new measure to compare networks, the Portrait Divergence, that is mathematically principled, incorporates the topological characteristics of networks at all structural scales, and is general-purpose and applicable to all types of networks. An important feature of our measure that enables many of its useful properties is that it is based on a graph invariant, the network portrait. We test our measure on both synthetic graphs and real world networks taken from protein interaction data, neuroscience, and computational social science applications. The Portrait Divergence reveals important characteristics of multilayer and temporal networks extracted from data.

3:00 – 4:00 Fika

Phil 4.14.18

Text Embedding Models Contain Bias. Here’s Why That Matters.

  • It occurs to me that bias may be a way of measuring societal dimension reduction. Need to read this carefully.
  • Neural network models can be quite powerful, effectively helping to identify patterns and uncover structure in a variety of different tasks, from language translation to pathology to playing games. At the same time, neural models (as well as other kinds of machine learning models) can contain problematic biases in many forms. For example, classifiers trained to detect rude, disrespectful, or unreasonable comments may be more likely to flag the sentence “I am gay” than “I am straight” [1]; face classification models may not perform as well for women of color [2]; speech transcription may have higher error rates for African Americans than White Americans [3].

Visual Analytics for Explainable Deep Learning

  • Recently, deep learning has been advancing the state of the art in artificial intelligence to a new level, and humans rely on artificial intelligence techniques more than ever. However, even with such unprecedented advancements, the lack of explanation regarding the decisions made by deep learning models and absence of control over their internal processes act as major drawbacks in critical decision-making processes, such as precision medicine and law enforcement. In response, efforts are being made to make deep learning interpretable and controllable by humans. In this paper, we review visual analytics, information visualization, and machine learning perspectives relevant to this aim, and discuss potential challenges and future research directions.

Submitted final version of the CI 2018 paper and also put a copy up on ArXive. Turns out that you can bundle everything into a tar file and upload once.

Phil 4.13.18

7:00 – ASRC MKT/BD

  • That Politico article on “news deserts” doesn’t really show what it claims to show
    • Its heart is in the right place, and the decline of local news really is a big threat to democratic governance.
  • Firing up the JuryRoom effort again
    • Unsurprisingly, there are updates
    • And a lot of fixing plugins. Big update
    • Ok, back to having PHP and MySQL working. Need to see how to integrate it with the Angular CLI
      • Updated CLI as per stackoverflow
        • In order to update the angular-cli package installed globally in your system, you need to run:

          npm uninstall -g angular-cli
          npm cache clean
          npm install -g @angular/cli@latest
          

          Depending on your system, you may need to prefix the above commands with sudo.

          Also, most likely you want to also update your local project version, because inside your project directory it will be selected with higher priority than the global one:

          rm -rf node_modules
          npm uninstall --save-dev angular-cli
          npm install --save-dev @angular/cli@latest
          npm install
          

          thanks grizzm0 for pointing this out on GitHub.

           

        • Updated my work environment too. Some PHP issues, and the Angular CLI wouldn’t update until I turned on the VPN. Duh.
      • Angular 4 + PHP: Setting Up Angular And Bootstrap – Part 2
    • Back to proposal writing

Phil 4.12.18

7:00 – 5:00 ASRC MKT/BD

  • Downloaded my FB DB today. Honestly, the only thing that seems excessive is the contact information
  • Interactive Semantic Alignment Model: Social Influence and Local Transmission Bottleneck
    • Dariusz Kalociński
    • Marcin Mostowski
    • Nina Gierasimczuk
    • We provide a computational model of semantic alignment among communicating agents constrained by social and cognitive pressures. We use our model to analyze the effects of social stratification and a local transmission bottleneck on the coordination of meaning in isolated dyads. The analysis suggests that the traditional approach to learning—understood as inferring prescribed meaning from observations—can be viewed as a special case of semantic alignment, manifesting itself in the behaviour of socially imbalanced dyads put under mild pressure of a local transmission bottleneck. Other parametrizations of the model yield different long-term effects, including lack of convergence or convergence on simple meanings only.
  • Starting to get back to the JuryRoom app. I need a better way to get the data parts up and running. This tutorial seems to have a minimal piece that works with PHP. That may be for the best since this looks like a solo effort for the foreseeable future
  • Proposal
    • Cut implementation down to proof-of-concept?
    • We are keeping the ASRC format
    • Got Dr. Lee’s contribution
    • And a lot of writing and figuring out of things

Phil 4.11.18

7:00 – 5:00 ASRC MKT

  • Fixed the quotes in Simon’s Anthill
  • Ordered Multiagent Systems: Algorithmic, Game-Theoretic, and Logical Foundations by Yoav Shoham.
  • Read more about SNM detection
  • Meeting with Aaron and T about aligning dev plan
  • More writing. We got a week extension!
    • Triaged exec summary
    • Triaged Transformational
  • Introducing TensorFlow Probability
    • At the 2018 TensorFlow Developer Summit, we announced TensorFlow Probability: a probabilistic programming toolbox for machine learning researchers and practitioners to quickly and reliably build sophisticated models that leverage state-of-the-art hardware. You should use TensorFlow Probability if:
      • You want to build a generative model of data, reasoning about its hidden processes.
      • You need to quantify the uncertainty in your predictions, as opposed to predicting a single value.
      • Your training set has a large number of features relative to the number of data points.
      • Your data is structured — for example, with groups, space, graphs, or language semantics — and you’d like to capture this structure with prior information.
      • You have an inverse problem — see this TFDS’18 talk for reconstructing fusion plasmas from measurements.
    • TensorFlow Probability gives you the tools to solve these problems. In addition, it inherits the strengths of TensorFlow such as automatic differentiation and the ability to scale performance across a variety of platforms: CPUs, GPUs, and TPUs.

Phil 4.10.18

7:00 – 5:00 ASRC MKT

  • Incorporating Wajanat’s changes
  • Discovered the csquotes package!
    \usepackage[autostyle]{csquotes}
    
    \begin{document}
    
    \enquote{Thanks!}
    
    \end{document}
  • Meeting with Drew
    • Nice chat. Basically, “use the databases!”
    • Also found this:
      • A Mechanism for Reasoning about Time and Belief
        • Hideki Isozaki
        • Yoav Shoham (Twitter)
        • Several computational frameworks have been proposed to maintain information about the evolving world, which embody a default persistence mechanism; examples include time maps and the event calculus. In multi-agent environments, time and belief both play essential roles. Belief interacts with time in two ways: there is the time at which something is believed, and the time about which it is believed. We augment the default mechanisms proposed for the purely temporal case so as to maintain information not only about the objective world but also about the evolution of beliefs. In the simplest case, this yields a two dimensional map of time, with persistence along each dimension. Since beliefs themselves may refer to other beliefs, we have to think of a statement referring to an agent’s temporal belief about another agent’s temporal belief ( a nested temporal belief statement). It poses both semantical and algorithmic problems. In this paper, we concentrate on the algorithmic aspect of the problems. The general case involves multi-dimensional maps of time called Temporal Belief Maps.
  • Register for CI 2018 – done
  • Finalize and submit paper by April 27, 2018
  • Did not get a go ahead for ONR
  • More work on the DHS proposal. Thinking about having a discussion about using latent values and clustering as the initial detection approach, and using ML as the initial simulation approach.
  • Then much banging away at keyboards. Good progress, I think
  • Neural Artistic Style Transfer: A Comprehensive Look

Phil 4.9.18

7:00 – ASRC MKT / BD

  • The Collective Intelligence 2018 paper was accepted! Now I need to start thinking about the presentation. And lodging, travel, etc.
  • Tweaking the SASO paper
  • The reasonably current version is on ArXive! Will update after submission to SASO this week.
  • This One Simple Trick Disrupts Digital Communities 
    • This paper describes an agent based simulation used to model human actions in belief space, a high-dimensional subset of information space associated with opinions. Using insights from animal collective behavior, we are able to simulate and identify behavior patterns that are similar to nomadic, flocking and stampeding patterns of animal groups. These behaviors have analogous manifestations in human interaction, emerging as solitary explorers, the fashion-conscious, and members of polarized echo chambers. We demonstrate that a small portion of nomadic agents that widely traverse belief space can disrupt a larger population of stampeding agents. Extending the model, we introduce the concept of Adversarial Herding, where bad actors can exploit properties of technologically mediated communication to artificially create self sustaining runaway polarization. We call this condition the Pishkin Effect as it recalls the large scale buffalo stampedes that could be created by native Americans hunters. We then discuss opportunities for system design that could leverage the ability to recognize these negative patterns, and discuss affordances that may disrupt the formation of natural and deliberate echo chambers.
  • Kind of between things, so I wrote up my notes on Influence of augmented humans in online interactions during voting events
  • Looks important: Lessons Learned Reproducing a Deep Reinforcement Learning Paper
  • Proposal all day today probably
  • Fika
  • add something about base model
  • echo chamber, bad actor

Phil 4.7.18

A Tale of Two Movements: Egypt During the Arab Spring and Occupy Wall Street

  • Social media provides flexible platforms that play key roles in energizing collective action in movements like Arab Spring (AS) and Occupy Wall Street (OWS). By enabling individuals to display emotions broadly, social media amplify sentiments defined as shared collective emotion to supply the forces that drive change in society. This study describes how one platform, Facebook, contributed to these two different examples of political activism. Using social network analytics and text mining, we examine how Fan Page posts during the life of the movements influenced the formation of social ties by using sentimental messaging. We hypothesize a set of relationships between group cohesion and polarity of sentiments in explaining involvement. We find that the strength of social ties formed through exchanges of posts and comments influence participation, but its effect differs across two movements. We also find that negative sentiments are associated with more participation for Egypt during the AS than OWS. Our results suggest cultural differences play a major role in participation behaviors. Social media is important in engineering management, because someone who has a negative reaction to a project or a product can use these media to reach thousands of individuals and potentially turn sentiment against a project.

Prefrontal cortex as a meta-reinforcement learning system

  • Over the past twenty years, neuroscience research on reward-based learning has converged on a canonical model, under which the neurotransmitter dopamine ‘stamps in’ associations between situations, actions and rewards by modulating the strength of synaptic connections between neurons. However, a growing number of recent findings have placed this standard model under strain. In the present work, we draw on recent advances in artificial intelligence to introduce a new theory of reward-based learning. Here, the dopamine system trains another part of the brain, the prefrontal cortex, to operate as its own free-standing learning system. This new perspective accommodates the findings that motivated the standard model, but also deals gracefully with a wider range of observations, providing a fresh foundation for future research.

Blade Runner And The Synthetic Panopticon

  •  There are already thousands of articles on misinformation, disinformation, and journalism flying by us every day in the US, in this very strange year, 2017. Rather than add to that, I simply intend to make several big picture observations that seem to be getting very little attention. Our present journalistic crisis comes to be not because people are merely misinformed about the truth, but because of a fundamental misunderstanding about how social power determines the construction of truth.

The disinformation order: Disruptive communication and the decline of democratic institutions

  • Many democratic nations are experiencing increased levels of false information circulating through social media and political websites that mimic journalism formats. In many cases, this disinformation is associated with the efforts of movements and parties on the radical right to mobilize supporters against centre parties and the mainstream press that carries their messages. The spread of disinformation can be traced to growing legitimacy problems in many democracies. Declining citizen confidence in institutions undermines the credibility of official information in the news and opens publics to alternative information sources. Those sources are often associated with both nationalist (primarily radical right) and foreign (commonly Russian) strategies to undermine institutional legitimacy and destabilize centre parties, governments and elections. The Brexit campaign in the United Kingdom and the election of Donald Trump in the United States are among the most prominent examples of disinformation campaigns intended to disrupt normal democratic order, but many other nations display signs of disinformation and democratic disruption. The origins of these problems and their implications for political communication research are explored.

Phil 4.6.18

7:00 – 9:00 ASRC MKT

  • Heard a San Francisco comedian refer to Google as “Mordor” to knowing laughter in the audience. That says a lot about the relationship between the SF folks and their technology nation-states to the south. It also makes me rethink what Mordor actually was…
  • More ArXive submission
    • Tips for submitting to ArXive for the first time
    • Make sure that only the used pix are uploaded
      • AdversarialHerding
      • EchoChamberAngle
      • Explore-Exploit
      • directionpreserving
      • SlewAngle
      • Explorer
      • coloredFlocking
      • stampede
      • RunawayTrace
      • populations
      • HerdingImpact
    • It may be possible to submit as a single zipped (.gz? .tar?)  package. Will try that next time
    • Submitted and pending approval.
  • Start on DHS proposal
    • Built LaTex document
    • The templates provided by ASRC are completely wrong. Fixed in the LaTex template
    • Lots of discussion and negotiation on the form of the concept. I think we’re ready to start Monday
  • Nice chat with Wajanat about the paper and then her work. It’s interesting to hear how references and metaphors that I think are common get missed when they are read by a non-native english speaker from a different cultural frame. For example, I refer to a “plague of locusts” , which I had to explain as one of the biblical plagues of Egypt. Once explained, Wajanat immediately got it, and mentioned the Arabic word طاعون, We then asked Ali, who’s Iranian. He didn’t know about plagues either, but by using طاعون, he was able to get the entire context. She also suggested improving the screenshot at the beginning of the paper and expanding the transition to the intelligent vehicle stampede section.
  • Then a meandering and fun chat with Shimei, mostly about psychology and AI ethics. Left at 9:00

Phil 4.5.18

7:00 – 5:00 ASRC MKT

  • More car stampedes: On one of L.A.’s steepest streets, an app-driven frenzy of spinouts, confusion and crashes
  • Working on the first draft of the paper. I think(?) I’m reasonably happy with it.
  • Trying to determine the submission guidelines. Are IEEE paper anonymized? If they are, here’s the post on how to do it and my implementation:
    \usepackage{xcolor}
    \usepackage{soul}
    
    \sethlcolor{black}
    \makeatletter
    \newif\if@blind
    \@blindfalse %use \@blindtrue to anonymize, \@blindfalse on final version
    \if@blind \sethlcolor{black}\else
    	\let\hl\relax
    \fi
    
    \begin{document}
    this text is \hl{redacted}
    \end{document}
    
    
  • So this clever solution doesn’t work, because you can select under the highlight. This is my much simpler solution:
    %\newcommand*{\ANON}{}
    \ifdefined\ANON
    	\author{\IEEEauthorblockN{Anonymous Author(s)}
    	\IEEEauthorblockA{\textit{this line kept for formatting} \\
    		\textit{this line kept for formatting}\\
    		this line kept for formatting \\
    		this line kept for formatting}
    }
    \else
    	\author{\IEEEauthorblockN{Philip Feldman}
    	\IEEEauthorblockA{\textit{ASRC Federal} \\
    	Columbia, USA \\
    	philip.feldman@asrcfederal.com}
    	}
    \fi
  • Submitting to Arxive
  • Boy, this hit home: The Swamp of Sadness
    • Even with Arteyu pulling on his bridle, Artex still had to start walking and keep walking to survive, and so do you. You have to pull yourself out of the swamp. This sucks, because it’s difficult, slow, hand-over-hand, gritty, horrible work, and you will end up very muddy. But I think the muddier the swamp, the better the learning really. I suspect the best kinds of teachers have themselves walked through very horrible swamps.
  • You have found the cui2vec explorer. This website will let you interact with embeddings for over 108,000 medical concepts. These embeddings were created using insurance claims for 60 million americans, 1.7 million full-text PubMed articles, and clinical notes from 20 million patients at Stanford. More information about the methods used to create these embeddings can be found in our preprint: https://arxiv.org/abs/1804.01486 
  • Going to James Foulds’ lecture on Mixed Membership Word Embeddings for Computational Social Science. Send email for meeting! Such JuryRoom! Done!
  • Kickoff meeting for the DHS proposal. We have until the 20th to write everything. Sheesh

Phil 4.4.18

7:00 – 5:00 ASRC MKT

  • From zero to research — An introduction to Meta-learning
    • Thomas Wolf Machine Learning, Natural Language Processing & Deep learning – Science Lead @ Huggingface  (We’re on a journey to build the first truly social artificial intelligence. Along the way, we contribute to the development of technology for the better.)
    • Over the last months, I have been playing and experimenting quite a lot with meta-learning models for Natural Language Processing and will be presenting some of this work at ICLR, next month in Vancouver 🇨🇦 — come say hi! 👋 In this post, I will start by making a very visual introduction to meta-learning, from zero to current research work. Then, we will code a meta-learning model in PyTorch from scratch and I will share some of the lessons learned on this project.

  • Google veteran Jeff Dean takes over as company’s AI chief
  • Add some MB framing words to the game theory part of the lit review – done
  • Work on the PSA writeup

Our research has indicated that an awareness of nomadic/explorer activity in belief space may help nudge stampeding groups away from a terminal trajectory and back towards “average” beliefs. Tajfel states that groups can exist “in opposition”, so providing counter-narratives may be ineffective. Rather, we think that a practical solution to online polarization is the injection of diversity into user’s feeds, be they social media, search results, videos, etc.  The infrastructure exists for this already in platform’s support of advertising. The precedent is the Public Service Announcement (PSA).

US Broadcasters since 1927, have been obligated to “serve the public interest” in exchange for spectrum rights. One way that this has been addressed is through the creation of the PSA, “the purpose of which is to improve the health, safety, welfare, or enhancement of people’s lives and the more effective and beneficial functioning of their community, state or region”

We believe that PSAs can be repurposed to support diversity injection through the following:

  • Random, non-political content designed to expand information horizons, analogous to clicking the “random article” link on Wikipedia.
  • Progressive levels of detail starting with an informative “hook” presented in social feeds or search results. Users should be able to explore as much or as little as they want.
  • Simultaneous presentation to large populations. Google has been approximating this with their “doodle” since 1998, with widespread positive feedback, which indicates that there may be good receptivity to common serendipitous information.
  • Format should reflect the medium, Text, images and videos.
  • Content should be easily verifiable, recognizable, and difficult to spoof.

We believe that such diversity injection mechanisms as described above can serve as a “first do no harm” first step in addressing the current crisis of misinformation. By nudging users towards an increased awareness of a wider world, which in turn interferes with the processes that lead to belief stampedes by increasing the number of dimensions, the awareness of different paths that others are taking. As we gain understanding of the mechanisms that influence group behaviors, it may be possible to further refine our designs and interfaces so that they no longer promote extremism while still providing value.

 

  • Done with first draft? Nope. Going to rework the implications section some more.

Phil 4.3.18

ASRC MKT 7:00 – 5:30

  • Integrating airplane notes on Influence of augmented humans in online interactions during voting events
  • Follow up on pointing logs
  • World Affairs Council (Part II. Part I is Jennifer Kavanagh and Tom Nichols: The End of Authority)
    • With so many forces undermining democratic institutions worldwide, we wanted a chance to take a step back and provide some perspective. Russian interference in elections here and in Europe, the rise in fake news and a decline in citizen trust worldwide pose a danger. In this second of a three part series, we look at the role of social media and the ways in which it was exploited for the purpose of sowing distrust. Janine Zacharia, former Jerusalem bureau chief and Middle East correspondent for The Washington Post, and Roger McNamee, managing director at Elevation Partners and an early stage investor in Google and Facebook, are in conversation with World Affairs CEO Jane Wales.
    • “The ultimate combination of propaganda and gambling … powered by machine learning”
  • The emergence of consensus: a primer (No Moscovici – odd)
    • The origin of population-scale coordination has puzzled philosophers and scientists for centuries. Recently, game theory, evolutionary approaches and complex systems science have provided quantitative insights on the mechanisms of social consensus. However, the literature is vast and widely scattered across fields, making it hard for the single researcher to navigate it. This short review aims to provide a compact overview of the main dimensions over which the debate has unfolded and to discuss some representative examples. It focuses on those situations in which consensus emerges ‘spontaneously’ in the absence of centralized institutions and covers topics that include the macroscopic consequences of the different microscopic rules of behavioural contagion, the role of social networks and the mechanisms that prevent the formation of a consensus or alter it after it has emerged. Special attention is devoted to the recent wave of experiments on the emergence of consensus in social systems.
  • Need to write up diversity injection proposal
    • Basically updated PSAs for social media
    • Intent is to expand the information horizon, not to counter anything in particular. So it’s not political
    • Presented in a variety of ways (maps, stories and lists)
    • Goes identically into everyone’s feed
    • Can be blocked, but blockers need to be studied
    • More injection as time on site goes up. Particularly with YouTube & FB
  • Working on SASO paper. Made it through discussion

Phil 4.2.18

7:00 – 5:00 ASRC MKT

  • Someone worked pretty hard on their April Fools joke
  • Started cleaning up my TF Dev Conf notes. Need to fill in speaker’s names and contacts – done
  • Contact Keith Bennet about “pointing” logs – done
  • Started editing the SASO flocking paper. Call is April 16!
    • Converted to LaTex and at 11 pages
  • But first – expense report…. Done! Forgot the parking though. Add tomorrow!
  • Four problems for news and democracy
    • To understand these four crises — addiction, economics, bad actors and known bugs — we have to look at how media has changed shape between the 1990s and today. A system that used to be linear and fairly predictable now features feedback loops that lead to complex and unintended consequences. The landscape that is emerging may be one no one completely understands, but it’s one that can be exploited even if not fully understood.
  • Humanitarianism’s other technology problem
    • Is social media affecting humanitarian crises and conflict in ways that kill people and may ultimately undermine humanitarian response?Fika. Meeting with Wajanat Friday to go over paper

     

Phil 3.30.18

TF Dev Sumit

Highlights blog post from the TF product manager

Keynote

  • Connecterra tracking cows
  • Google is an AI – first company. All products are being influenced. TF is the dogfood that everyone is eating at google.

Rajat Monga

  • Last year has been focussed on making TF easy to use
  • 11 million downloads
  • blog.tensorflow.org
  • youtube.com/tensorflow
  • tensorflow.org/ub
  • tf.keras – full implementation.
  • Premade estimators
  • three line training from reading to model? What data formats?
  • Swift and tensorflow.js

Megan

  • Real-world data and time-to-accuracy
  • Fast version is the pretty version
  • TensorflowLite is 300% speedup in inference? Just on mobile(?)
  • Training speedup is about 300% – 400% anually
  • Cloud TPUs are available in V2. 180 TF computation
  • github.com/tensorflow/tpu
  • ResNet-50 on Cloud TPU in < 15

Jeff Dean

  • Grand Engineering challenges as a list of  ML goals
  • Engineer the tools for scientific discovery
  • AutoML – Hyperparameter tuning
  • Less expertise (What about data cleaning?)
    • Neural architecture search
    • Cloud Automl for computer vision (for now – more later)
  • Retinal data is being improved as the data labeling improves. The trained human trains the system proportionally
  • Completely new, novel scientific discoveries – machine scan explore horizons in different ways from humans
  • Single shot detector

Derrek Murray @mrry (tf.data)

  • Core TF team
  • tf.data  –
  • Fast, Flexible, and Easy to use
    • ETL for TF
    • tensorflow.org/performance/datasets_performance
    • Dataset tf.SparseTensor
    • Dataset.from_generator – generates graphs from numpy arrays
    • for batch in dataset: train_model(batch)
    • 1.8 will read in CSV
    • tf.contrib.data.make_batched_features_dataset
    • tf.contrib.data.make_csv_dataset()
    • Figures out types from column names

Alexandre Passos (Eager Execution)

  • Eager Execution
  • Automatic differentiation
  • Differentiation of graphs and code <- what does this mean?
  • Quick iterations without building graphs
  • Deep inspection of running models
  • Dynamic models with complex control flows
  • tf.enable_eager_execution()
  • immediately run the tf code that can then be conditional
  • w = tfe.variables([[1.0]])
  • tape to record actions, so it’s possible to evaluate a variety of approaches as functions
  • eager supports debugging!!!
  • And profilable…
  • Google collaboratory for Jupyter
  • Customizing gradient, clipping to keep from exploding, etc
  • tf variables are just python objects.
  • tfe.metrics
  • Object oriented savings of TF models Kind of like pickle, in that associated variables are saved as well
  • Supports component reuse?
  • Single GPU is competitive in speed
  • Interacting with graphs: Call into graphs Also call into eager from a graph
  • Use tf.keras.layers, tf.keras.Model, tf.contribs.summary, tfe.metrics, and object-based saving
  • Recursive RNNs work well in this
  • Live demo goo.gl/eRpP8j
  • getting started guide tensorflow.org/programmers_guide/eager
  • example models goo.gl/RTHJa5

Daniel Smilkov (@dsmilkov) Nikhl Thorat (@nsthorat)

  • In-Browser ML (No drivers, no installs)
  • Interactive
  • Browsers have access to sensors
  • Data stays on the client (preprocessing stage)
  • Allows inference and training entirely in the browser
  • Tensorflow.js
    • Author models directly in the browser
    • import pre-trained models for inference
    • re-train imported models (with private data)
    • Layers API, (Eager) Ops API
    • Can port keras or TF morel
  • Can continue to train a model that is downloaded from the website
  • This is really nice for accessibility
  • js.tensorflow.org
  • github.com/tensorflow/tfjs
  • Mailing list: goo.gl/drqpT5

Brennen Saeta

  • Performance optimization
  • Need to be able to increase performance exponentially to be able to train better
  • tf.data is the way to load data
  • Tensorboard profiling tools
  • Trace viewer within Tensorboard
  • Map functions seem to take a long time?
  • dataset.map(Parser_fn, num_parallel_calls = 64)) <- multithreading
  • Software pipelining
  • Distributed datasets are becoming critical. They will not fit on a single instance
  • Accelerators work in a variety of ways, so optimizing is hardware dependent For example, lower precision can be much faster
  • bfloat16 brain floating point format. Better for vanishing and exploding gradients
  • Systolic processors load the hardware matrix while it’s multiplying, since you start at the upper left corner…
  • Hardware is becoming harder and harder to do apples-to apples. You need to measure end-to-end on your own workloads. As a proxy, Stanford’s DAWNBench
  • Two frameworks XLA nd Graph

Mustafa Ispir (tf.estimator, high level modules for experiments and scaling)

  • estimators fill in the model, based on Google experiences
  • define as an ml problem
  • pre made estimators
  • reasonable defaults
  • feature columns – bucketing, embedding, etc
  • estimator = model_to_estimator
  • image = hum.image_embedding_column(…)
  • supports scaling
  • export to production
  • estimator.export_savemodel()
  • Feature columns (from csv, etc) intro, goo.gl/nMEPBy
  • Estimators documentation, custom estimators
  • Wide-n-deep (goo.gl/l1cL3N from 2017)
  • Estimators and Keras (goo.gl/ito9LE Effective TensorFlow for Non-Experts)

Igor Sapirkin

  • distributed tensorflow
  • estimator is TFs highest level of abstraction in the API google recommends using the highest level of abstraction you can be effective in
  • Justine debugging with Tensorflow Debugger
  • plugins are how you add features
  • embedding projector with interactive label editing

Sarah Sirajuddin, Andrew Selle (TensorFlow Lite) On-device ML

  • TF Lite interpreter is only 75 kilobytes!
  • Would be useful as a biometric anonymizer for trustworthy anonymous citizen journalism. Maybe even adversarial recognition
  • Introduction to TensorFlow Lite → https://goo.gl/8GsJVL
  • Take a look at this article “Using TensorFlow Lite on Android” → https://goo.gl/J1ZDqm

Vijay Vasudevan AutoML @spezzer

  • Theory lags practice in valuable discipline
  • Iteration using human input
  • Design your code to be tunable at all levels
  • Submit your idea to an idea bank

Ian Langmore

  • Nuclear Fusion
  • TF for math, not ML

Cory McLain

  • Genomics
  • Would this be useful for genetic algorithms as well?

Ed Wilder-James

  • Open source TF community
  • Developers mailing list developers@tensorflow.org
  • tensorflow.org/community
  • SIGs SIGBuild, other coming up
  • SIG Tensorboard <- this

Chris Lattner

  • Improved usability of TF
  • 2 approaches, Graph and Eager
  • Compiler analysis?
  • Swift language support as a better option than Python?
  • Richard Wei
  • Did not actually see the compilation process with error messages?

TensorFlow Hub Andrew Gasparovic and Jeremiah Harmsen

  • Version control for ML
  • Reusable module within the hub. Less than a model, but shareable
  • Retrainable and backpropagateable
  • Re-use the architecture and trained weights (And save, many, many, many hours in training)
  • tensorflow.org/hub
  • module = hub.Module(…., trainable = true)
  • Pretrained and ready to use for classification
  • Packages the graph and the data
  • Universal Sentence Encodings semantic similarity, etc. Very little training data
  • Lower the learning rate so that you don’t ruin the existing rates
  • tfhub.dev
  • modules are immutable
  • Colab notebooks
  • use #tfhub when modules are completed
  • Try out the end-to-end example on GitHub → https://goo.gl/4DBvX7

TF Extensions Clemens Mewald and Raz Mathias

  • TFX is developed to support lifecycle from data gathering to production
  • Transform: Develop training model and serving model during development
  • Model takes a raw data model as the request. The transform is being done in the graph
  • RESTful API
  • Model Analysis:
  • ml-fairness.com – ROC curve for every group of users
  • github.com/tensorflow/transform

Project Magenta (Sherol Chen)

People:

  • Suharsh Sivakumar – Google
  • Billy Lamberta (documentation?) Google
  • Ashay Agrawal Google
  • Rajesh Anantharaman Cray
  • Amanda Casari Concur Labs
  • Gary Engler Elemental Path
  • Keith J Bennett (bennett@bennettresearchtech.com – ask about rover decision transcripts)
  • Sandeep N. Gupta (sandeepngupta@google.com – ask about integration of latent variables into TF usage as a way of understanding the space better)
  • Charlie Costello (charlie.costello@cloudminds.com – human robot interaction communities)
  • Kevin A. Shaw (kevin@algoint.com data from elderly to infer condition)

 

Phil 3.28.18

7:00 – 5:00 ASRC MKT

    • Aaron found this hyperparameter optimization service: Sigopt
      • Improve ML models 100x faster
      • SigOpt’s API tunes your model’s parameters through state-of-the-art Bayesian optimization.
      • Exponentially faster and more accurate than grid search. Faster, more stable, and easier to use than open source solutions.
      • Extracts additional revenue and performance left on the table by conventional tuning.
    • A Strategy for Ranking Optimization Methods using Multiple Criteria
      • An important component of a suitably automated machine learning process is the automation of the model selection which often contains some optimal selection of hyperparameters. The hyperparameter optimization process is often conducted with a black-box tool, but, because different tools may perform better in different circumstances, automating the machine learning workflow might involve choosing the appropriate optimization method for a given situation. This paper proposes a mechanism for comparing the performance of multiple optimization methods for multiple performance metrics across a range of optimization problems. Using nonparametric statistical tests to convert the metrics recorded for each problem into a partial ranking of optimization methods, results from each problem are then amalgamated through a voting mechanism to generate a final score for each optimization method. Mathematical analysis is provided to motivate decisions within this strategy, and sample results are provided to demonstrate the impact of certain ranking decisions
    • World Models: Can agents learn inside of their own dreams?
      • We explore building generative neural network models of popular reinforcement learning environments[1]. Our world model can be trained quickly in an unsupervised manner to learn a compressed spatial and temporal representation of the environment. By using features extracted from the world model as inputs to an agent, we can train a very compact and simple policy that can solve the required task. We can even train our agent entirely inside of its own hallucinated dream generated by its world model, and transfer this policy back into the actual environment.
    • Tweaked the SingleNeuron spreadsheet
    • This came up again: A new optimizer using particle swarm theory (1995)
      • The optimization of nonlinear functions using particle swarm methodology is described. Implementations of two paradigms are discussed and compared, including a recently developed locally oriented paradigm. Benchmark testing of both paradigms is described, and applications, including neural network training and robot task learning, are proposed. Relationships between particle swarm optimization and both artificial life and evolutionary computation are reviewed.
      • New: Particle swarm optimization for hyper-parameter selection in deep neural networks
    • Working with the CIFAR10 data now. Tradeoff between filters and epochs:
      NB_EPOCH = 10
      NUM_FIRST_FILTERS = int(32/2)
      NUM_MIDDLE_FILTERS = int(64/2)
      OUTPUT_NEURONS = int(512/2)
      Test score: 0.8670728429794311
      Test accuracy: 0.6972
      Elapsed time =  565.9446044602014
      
      NB_EPOCH = 5
      NUM_FIRST_FILTERS = int(32/1)
      NUM_MIDDLE_FILTERS = int(64/1)
      OUTPUT_NEURONS = int(512/1)
      Test score: 0.8821897733688354
      Test accuracy: 0.6849
      Elapsed time =  514.1915690121759
      
      NB_EPOCH = 10
      NUM_FIRST_FILTERS = int(32/1)
      NUM_MIDDLE_FILTERS = int(64/1)
      OUTPUT_NEURONS = int(512/1)
      Test score: 0.7007060846328735
      Test accuracy: 0.765
      Elapsed time =  1017.0974014300725
      
      Augmented imagery
      NB_EPOCH = 10
      NUM_FIRST_FILTERS = int(32/1)
      NUM_MIDDLE_FILTERS = int(64/1)
      OUTPUT_NEURONS = int(512/1)
      Test score: 0.7243581249237061
      Test accuracy: 0.7514
      Elapsed time =  1145.673343808471
      
    • And yet, something is clearly wrong: wrongPNG
    • Maybe try this version? samyzaf.com/ML/cifar10/cifar10.html