Category Archives: Simulation

Phil 10.31.18

7:00 – ASRC PhD

  • Read this carefully today: Introducing AdaNet: Fast and Flexible AutoML with Learning Guarantees
    • Today, we’re excited to share AdaNet, a lightweight TensorFlow-based framework for automatically learning high-quality models with minimal expert intervention. AdaNet builds on our recent reinforcement learning and evolutionary-based AutoML efforts to be fast and flexible while providing learning guarantees. Importantly, AdaNet provides a general framework for not only learning a neural network architecture, but also for learning to ensemble to obtain even better models.
    • What about data from simulation?
    • Github repo
    • This looks like it’s based deeply the cloud AI and Machine Learning products, including cloud-based hyperparameter tuning.
    • Time series prediction is here as well, though treated in a more BigQuery manner
      • In this blog post we show how to build a forecast-generating model using TensorFlow’s DNNRegressor class. The objective of the model is the following: Given FX rates in the last 10 minutes, predict FX rate one minute later.
    • Text generation:
      • Cloud poetry: training and hyperparameter tuning custom text models on Cloud ML Engine
        • Let’s say we want to train a machine learning model to complete poems. Given one line of verse, the model should generate the next line. This is a hard problem—poetry is a sophisticated form of composition and wordplay. It seems harder than translation because there is no one-to-one relationship between the input (first line of a poem) and the output (the second line of the poem). It is somewhat similar to a model that provides answers to questions, except that we’re asking the model to be a lot more creative.
      • Codelab: Google Developers Codelabs provide a guided, tutorial, hands-on coding experience. Most codelabs will step you through the process of building a small application, or adding a new feature to an existing application. They cover a wide range of topics such as Android Wear, Google Compute Engine, Project Tango, and Google APIs on iOS.
        Codelab tools on GitHub

  • Add the Range and Length section in my notes to the DARPA measurement section. Done. I need to start putting together the dissertation using these parts
  • Read Open Source, Open Science, and the Replication Crisis in HCI. Broadly, it seems true, but trying to piggyback on GitHub seems like a shallow solution that repurposes something for coding – an ephemeral activity, to science, which is archival for a reason. Thought needs to be given to an integrated (collection, raw data, cleaned data, analysis, raw results, paper (with reviews?), slides, and possibly a recording of the talk with questions. What would it take to make this work across all science, from critical ethnographies to particle physics? How will it be accessible in 100 years? 500? 1,000? This is very much an HCI problem. It is about designing a useful socio-cultural interface. Some really good questions would be “how do we use our HCI tools to solve this problem?”, and, “does this point out the need for new/different tools?”.
  • NASA AIMS meeting. Demo in 2 weeks. AIMS is “time series prediction”, A2P is “unstructured data”. Proove that we can actually do ML, as opposed to saying things.
    • How about cross-point correlation? Could show in a sim?
    • Meeting on Friday with a package
    • We’ve solved A, here’s the vision for B – Z and a roadmap. JPSS is a near-term customer (JPSS Data)
    • Getting actionable intelligence from the system logs
    • Application portfolios for machine learning
    • Umbrella of capabilities for Rich Burns
    • New architectural framework for TTNC
    • Complete situational awareness. Access to commands and sensor streams
    • Software Engineering Division/Code 580
    • A2P as a toolbox, but needs to have NASA-relevant analytic capabilities
    • GMSEC overview

Phil 10.30.18

7:00 – 3:30 ASRC PhD

  • Search as embodies in the “Ten Blue Links” meets the requirements of a Parrow “Normal Accident”
    • The search results are densely connected. That’s how PageRank works. Even latent connections matter.
    • The change in popularity of a page rapidly affects the rank. So the connections are stiff
    • The relationships of the returned links both to each other and to the broader information landscape in general is hidden.
    • An additional density and stiffness issue is that everyone uses Google, so there is a dense, stiff connection between the search engine and the population of users
  • Write up something about how
    • ML can make maps, which decrease the likelihood of IR contributing to normal accidents
    • AI can use these maps to understand the shape of human belief space, and where the positive regions and dangerous sinks are.
  • Two measures for maps are the concepts or Range and length. Range is the distance that a trajectory can be placed on the map and remain contiguous. Length is the total distance that a trajectory travels, independent of the map its placed on.
  • Write up the basic algorithm of ML to map production
    • Take a set of trajectories that are known to be in the same belief region (why JuryRoom is needed) as the input
    • Generate an N-dimensional coordinate frame that best preserves length over the greatest range.
    • What is used as the basis for the trajectory may matter. The range (at a minimum), can go from letters to high-level topics. I think any map reconstruction based on letters would be a tangle, with clumps around TH, ER, ON, and AN. At the other end, an all-encompassing meta-topic, like WORDS would be a single, accurate, but useless single point. So the map reconstruction will become possible somewhere between these two extremes.
  • The Nietzsche text is pretty good. In particular, check out the way the sentences form based on the seed  “s when one is being cursed.
    • the fact that the spirit of the spirit of the body and still the stands of the world
    • the fact that the last is a prostion of the conceal the investion, there is our grust
    • the fact them strongests! it is incoke when it is liuderan of human particiay
    • the fact that she could as eudop bkems to overcore and dogmofuld
    • In this case, the first 2-3 words are the same, and random, semi-structured text. That’s promising, since the compare would be on the seed plus the generated text.
  • Today, see how fast a “Shining” (All work and no play makes Jack a dull boy.) text can be learned and then try each keyword as a start. As we move through the sentence, the probability of the next words should change.
    • Generate the text set
    • Train the Nietzsche model on the new text. Done. Here are examples with one epoch and a batch size of 32, with a temperature of 1.0:
      ----- diversity: 0.2
      ----- Generating with seed: "es jack a 
      dull boy all work and no play"
      es jack a 
      dull boy all work and no play makes jack a dull boy all work and no play makes jack a dull boy all work and no play makes jack a dull boy all work and no play makes 
      
      ----- diversity: 0.5
      ----- Generating with seed: "es jack a 
      dull boy all work and no play"
      es jack a 
      dull boy all work and no play makes jack a dull boy all work and no play makes jack a dull boy all work and no play makes jack a dull boy all work and no play makes 
      
      ----- diversity: 1.0
      ----- Generating with seed: "es jack a 
      dull boy all work and no play"
      es jack a 
      dull boy all work and no play makes jack a dull boy anl wory and no play makes jand no play makes jack a dull boy all work and no play makes jack a 
      
      ----- diversity: 1.2
      ----- Generating with seed: "es jack a 
      dull boy all work and no play"
      es jack a 
      dull boy all work and no play makes jack a pull boy all work and no play makes jack andull boy all work and no play makes jack a dull work and no play makes jack andull

      Note that the errors start with a temperature of 1.0 or greater

    • Rewrite the last part of the code to generate text based on each word in the sentence.
      • So I tried that and got gobbledygook. The issues is that the prediction only works on waveform-sized chunks. To verify this, I created a seed from the input text, truncating it to maxlen (20 in this case):
        sentence = "all work and no play makes jack a dull boy"[:maxlen]

        That worked, but it means that the character-based approach isn’t going to work

        ----- temperature: 0.2
        ----- Generating with seed: [all work and no play]
        all work and no play makes jack a dull boy all work and no play makes jack a dull boy all work and no play makes 
        
        ----- temperature: 0.5
        ----- Generating with seed: [all work and no play]
        all work and no play makes jack a dull boy all work and no play makes jack a dull boy all work and no play makes 
        
        ----- temperature: 1.0
        ----- Generating with seed: [all work and no play]
        all work and no play makes jack a dull boy all work and no play makes jack a dull boy pllwwork wnd no play makes 
        
        ----- temperature: 1.2
        ----- Generating with seed: [all work and no play]
        all work and no play makes jack a dull boy all work and no play makes jack a dull boy all work and no play makes

         

    • Based on this result and the ensuing chat with Aaron, we’re going to revisit the whole LSTM with numbers and build out a process that will support words instead of characters.
  • Looking for CMAC models, I found Self Organizing Feature Maps at NeuPy.com:
  • Here’s How Much Bots Drive Conversation During News Events
    • Late last week, about 60 percent of the conversation was driven by likely bots. Over the weekend, even as the conversation about the caravan was overshadowed by more recent tragedies, bots were still driving nearly 40 percent of the caravan conversation on Twitter. That’s according to an assessment by Robhat Labs, a startup founded by two UC Berkeley students that builds tools to detect bots online. The team’s first product, a Chrome extension called BotCheck.me, allows users to see which accounts in their Twitter timelines are most likely bots. Now it’s launching a new tool aimed at news organizations called FactCheck.me, which allows journalists to see how much bot activity there is across an entire topic or hashtag

Phil 10.29.18

7:00 – 5:00 ASRC PhD

  • This looks like a Big Deal from Google – Working together to apply AI for social good
    • Google.org is issuing an open call to organizations around the world to submit their ideas for how they could use AI to help address societal challenges. Selected organizations will receive support from Google’s AI experts, Google.org grant funding from a $25M pool, credit and consulting from Google Cloud, and more.
    • We look forward to receiving your application on or before 11:59 p.m. PT on January 22, 2019, and we encourage you to apply early given that we expect high volume within the last few hours of the application window. Thank you!
    • Application Guide
    • Application form (can’t save, compose offline using guide, above)
  • Finished my writeup on Meltdown
  • Waiting for a response from Antonio
  • Meeting with Don at 9:00 to discuss BAA partnership.
    • Don is comfortable with being PI or co-PI, whichever works best. When we call technical POCs, we speak on his behalf
    • We discussed how he could participate with the development of theoretical models based on signed graph Laplacians creating structures that can move in belief space. He thinks the idea has merit, and can put in up to 30% of his time on mathematical models and writing
    • ASRC has already partnered with UMBC. ASRC would sub to UMBC
    • Ordinarily, IP is distributed proportional to the charged hours
    • Don has access to other funding vehicles that can support the Army BAA, but this would make things more complicated. These should be discussed if we can’t make a “clean” agreement that meets our funding needs
  • Pinged Brian about his defense.
  • Some weekend thoughts
    • Opinion dynamics systems describe how communication within a network occurs, but disregards the motion of the network as a whole. In cases when the opinions converge, the network is stiff.
    • Graph laplacians could model “othering” by having negative weights. It looks like these are known as signed laplacians, and useful to denote difference. The trick is to discover the equations of motion. How do you model a “social particle”?
  • Just discovered the journal Swarm Intelligence
    • Swarm Intelligence is the principal peer reviewed publication dedicated to reporting research and new developments in this multidisciplinary field. The journal publishes original research articles and occasional reviews on theoretical, experimental, and practical aspects of swarm intelligence. It offers readers reports on advances in the understanding and utilization of systems that are based on the principles of swarm intelligence. Emphasis is given to such topics as the modeling and analysis of collective biological systems; application of biological swarm intelligence models to real-world problems; and theoretical and empirical research in ant colony optimization, particle swarm optimization, swarm robotics, and other swarm intelligence algorithms. Articles often combine experimental and theoretical work.
  • I think it’s time to start ramping up on the text generation!
      • Updated my home box to tensorflow 1.11.0. Testing to see if it still works using the Deep Learning with Keras simple_nueral_net.py example. Hasn’t broken (yet…), but is taking a long time… Worked! And it’s much faster the second time. Don’t know why that is and can’t find anything online that talks to that.
        Loss: 0.5043802047491074
        Accuracy: 0.8782
        Time =  211.42629722093085
      • Found this keras example for generating Nietsche

     

    • Trying it out. This may be a overnight run… But it is running.
  • Had a good discussion with Aaron about how mapmaking could be framed as an ML problem. More writeup tomorrow.

Phil 10.17.18

7:00 – 4:00 Antonio Workshop

Phil 10.8.18

7:00 – 12:00, 2:00 – 5:00 ASRC Research

  • Finish up At Home in the Universe notes – done!
  • Get started on framing out Antonio’s paper – good progress!
    • Basically, Aaron and I think there is a spectrum of interaction that can occur in these systems. At one end is some kind of market, where communication is mediated through price, time, and convenience to the transportation user. At the other is a more top down, control system way of dealing with this. NIST RCS would be an example of this. In between these two extremes are control hierarchies that in turn interact through markets
  • Wrote up some early thoughts on how simulation and machine learning can be a thinking fast and slow solution to understandable AI

Phil 9.28.18

7:30 – 4:00 ASRC MKT

  • Stumbled on this podcast this morning: How Small Problems Snowball Into Big Disasters
  • How to Prepare for a Crisis You Couldn’t Possibly Predict
  • I’m trying to think about how this should be applied to human/machine ecologies. I think that simulation is really important because it lets one model patch compare itself against another model without real-world impacts. This has something to do with a shared, multi-instance environment simulation as well. The environment provides one level of transparent interaction, but there also needs to be some level of inadvertent social information that shows some insight into how a particular system is working.
    • When the simulation and the real world start to diverge for a system, that needs to be signaled
    • Systems need to be able to “look into” other simulations and compare like with like. So a tagged item (bicycle) in one sim is the same in another.
    • Is there an OS that hands out environments?
    • How does a decentralized system coordinate? Is there an answer in MMOGs?
  • Kate Starbird’s presentation was interesting as always. We had a chance to talk afterwards, and she’d like to see our work, so I’ve sent her links to the last two papers.
    I also met Bill Braniff, who is the director of the UMD Study of Terrorism and responses to Terrorism. He got papers too, with a brief description about how mapping could aid in the detection of radicalization patterns
    Then at lunch, I had a chance to meet with Roger Bostelman from NIST. He’s interested in writing standards for fleet and swarm vehicles, and is interested in making sure that standards mitigate the chance of stampeding autonomous vehicles, so I sent him the Blue Sky draft.
    And lastly, I got a phone call from Aaron who says that our project will be terminated December 31, after which there will be no more IR&D at ASRC. It was a nice run while it lasted. And they may change their minds, but I doubt it.

Phil 9.21.18

7:00 – 4:00 ASRC MKT

  • “Who’s idea was it to connect every idiot on the internet with every other idiot” PJ O’Rourke, Commonwealth Club, 2018
  • Running Programs In Reverse for Deeper A.I.” by Zenna Tavares
    • In this talk I show that inverse simulation, i.e., running programs in reverse from output to input, lies at the heart of the hardest problems in both human cognition and artificial intelligence. How humans are able to reconstruct the rich 3D structure of the world from 2D images; how we predict that it is safe to cross a street just by watching others walk, and even how we play, and sometimes win at Jenga, are all solvable by running programs backwards. The idea of program inversion is old, but I will present one of the first approaches to take it literally. Our tool ReverseFlow combines deep-learning and our theory of parametric inversion to compile the source code of a program (e.g., a TensorFlow graph) into its inverse, even when it is not conventionally invertible. This framework offers a unified and practical approach to both understand and solve the aforementioned problems in vision, planning and inference for both humans and machines.
  • Bot-ivistm: Assessing Information Manipulation in Social Media Using Network Analytics
    • Matthew Benigni 
    • Kenneth Joseph
    • Kathleen M. Carley (Scholar)
    • Social influence bot networks are used to effect discussions in social media. While traditional social network methods have been used in assessing social media data, they are insufficient to identify and characterize social influence bots, the networks in which they reside and their behavior. However, these bots can be identified, their prevalence assessed, and their impact on groups assessed using high dimensional network analytics. This is illustrated using data from three different activist communities on Twitter—the “alt-right,” ISIS sympathizers in the Syrian revolution, and activists of the Euromaidan movement. We observe a new kind of behavior that social influence bots engage in—repetitive @mentions of each other. This behavior is used to manipulate complex network metrics, artificially inflating the influence of particular users and specific agendas. We show that this bot behavior can affect network measures by as much as 60% for accounts that are promoted by these bots. This requires a new method to differentiate “promoted accounts” from actual influencers. We present this method. We also present a method to identify social influence bot “sub-communities.” We show how an array of sub-communities across our datasets are used to promote different agendas, from more traditional foci (e.g., influence marketing) to more nefarious goals (e.g., promoting particular political ideologies).
  • Pinged Aaron M. about writing an article
  • More iConf paper. Got a first draft on everything but the discussion section

Phil 8.30.18

7:00 – 5:00  ASRC MKT

  • Target Blue Sky paper for iSchool/iConference 2019: The chairs are particularly looking for “Blue Sky Ideas” that are open-ended, possibly even “outrageous” or “wacky,” and present new problems, new application domains, or new methodologies that are likely to stimulate significant new research. 
  • I’m thinking that a paper that works through the ramifications of this diagram as it relates to people and machines. With humans that are slow responding with spongy, switched networks the flocking area is large. With a monolithic densely connected system it’s going to be a straight line from nomadic to stampede. Nomad-Flocking-Stampede2
    • Length: Up to 4 pages (excluding references)
    • Submission deadline: October 1, 2018
    • Notification date: mid-November, 2018
    • Final versions due: December 14, 2018
    • First versions will be submitted using .pdf. Final versions must be submitted in .doc, .docx or La Tex.
  • More good stuff on BBC Business Daily Trolling for Cash
    • Anger and animosity is prevalent online, with some people even seeking it out. It’s present on social media of course as well as many online forums. But now outrage has spread to mainstream media outlets and even the advertising industry. So why is it so lucrative? Bonny Brooks, a writer and researcher at Newcastle University explains who is making money from outrage. Neuroscientist Dr Dean Burnett describes what happens to our brains when we see a comment designed to provoke us. And Curtis Silver, a tech writer for KnowTechie and ForbesTech, gives his thoughts on what we need to do to defend ourselves from this onslaught of outrage.
  • Exposure to Opposing Views can Increase Political Polarization: Evidence from a Large-Scale Field Experiment on Social Media
    • Christopher Bail (Scholar)
    • There is mounting concern that social media sites contribute to political polarization by creating “echo chambers” that insulate people from opposing views about current events. We surveyed a large sample of Democrats and Republicans who visit Twitter at least three times each week about a range of social policy issues. One week later, we randomly assigned respondents to a treatment condition in which they were offered financial incentives to follow a Twitter bot for one month that exposed them to messages produced by elected officials, organizations, and other opinion leaders with opposing political ideologies. Respondents were re-surveyed at the end of the month to measure the effect of this treatment, and at regular intervals throughout the study period to monitor treatment compliance. We find that Republicans who followed a liberal Twitter bot became substantially more conservative post-treatment, and Democrats who followed a conservative Twitter bot became slightly more liberal post-treatment. These findings have important implications for the interdisciplinary literature on political polarization as well as the emerging field of computational social science.
  • Setup gcloud tools on laptop – done
  • Setup Tensorflow on laptop. Gave up un using CUDA 9.1, but got tf doing ‘hello, tensorflow’
  • Marcom meeting – 2:00
  • Get the concept of behaviors being a more scalable, dependable way of vetting information.
    • Eg Watching the DISI of outrage as manifested in trolling
      • “Uh. . . . not to be nitpicky,,,,,but…the past tense of drag is dragged, not drug.”: An overview of trolling strategies
        • Dr Claire Hardaker (Scholar) (Blog)
          • I primarily research aggression, deception, and manipulation in computer-mediated communication (CMC), including phenomena such as flaming, trolling, cyberbullying, and online grooming. I tend to take a forensic linguistic approach, based on a corpus linguistic methodology, but due to the multidisciplinary nature of my research, I also inevitably branch out into areas such as psychology, law, and computer science.
        • This paper investigates the phenomenon known as trolling — the behaviour of being deliberately antagonistic or offensive via computer-mediated communication (CMC), typically for amusement’s sake. Having previously started to answer the question, what is trolling? (Hardaker 2010), this paper seeks to answer the next question, how is trolling carried out? To do this, I use software to extract 3,727 examples of user discussions and accusations of trolling from an eighty-six million word Usenet corpus. Initial findings suggest that trolling is perceived to broadly fall across a cline with covert strategies and overt strategies at each pole. I create a working taxonomy of perceived strategies that occur at different points along this cline, and conclude by refining my trolling definition.
        • Citing papers
  • FireAnt (Filter, Identify, Report, and Export Analysis Toolkit) is a freeware social media and data analysis toolkit with built-in visualization tools including time-series, geo-position (map), and network (graph) plotting.
  • Fix marquee – done
  • Export to ppt – done!
    • include videos – done
    • Center title in ppt:
      • model considerations – done
      • diversity injection – done
  • Got the laptop running Python and Tensorflow. Had a stupid problem where I accidentally made a virtual environment and keras wouldn’t work. Removed, re-connected and restarted IntelliJ and everything is working!

Phil 8.19.18

7:00 – 5:30 ASRC MKT

  • Had a thought that the incomprehension that comes from misalignment that Stephens shows resembles polarizing light. I need to add a slider that enables influence as a function of alignment. Done
    • Getting the direction cosine between the source and target belief
      double interAgentDotProduct = unitOrientVector.dotProduct(otherUnitOrientVector);
      double cosTheta = Math.min(1.0, interAgentDotProduct);
      double beliefAlignment = Math.toDegrees(Math.acos(cosTheta));
      double interAgentAlignment = (1.0 - beliefAlignment/180.0);
    • Adding a global variable that sets how much influence (0% – 100%) influence from an opposing agent. Just setting it to on/off, because the effects are actually pretty subtle
  • Add David’s contributions to slide one writeup – done
  • Start slide 2 writeup
  • Find casters for Dad’s walker
  • Submit forms for DME repair
    • Drat – I need the ECU number
  • Practice talk!
    • Need to reduce complexity and add clearly labeled sections, in particular methods
  • I need to start paying attention to attention
  • Also, keeping this on the list How social media took us from Tahrir Square to Donald Trump by Zeynep Tufekci
  • Social Identity Threat Motivates Science – Discrediting Online Comments
    • Experiencing social identity threat from scientific findings can lead people to cognitively devalue the respective findings. Three studies examined whether potentially threatening scientific findings motivate group members to take action against the respective findings by publicly discrediting them on the Web. Results show that strongly (vs. weakly) identified group members (i.e., people who identified as “gamers”) were particularly likely to discredit social identity threatening findings publicly (i.e., studies that found an effect of playing violent video games on aggression). A content analytical evaluation of online comments revealed that social identification specifically predicted critiques of the methodology employed in potentially threatening, but not in non-threatening research (Study 2). Furthermore, when participants were collectively (vs. self-) affirmed, identification did no longer predict discrediting posting behavior (Study 3). These findings contribute to the understanding of the formation of online collective action and add to the burgeoning literature on the question why certain scientific findings sometimes face a broad public opposition.

Phil 7.18.18

divylmzuyaeqjbk

There was no colusion“…”Anyone involved in that meddling to justice.

Premises for Data Science Magical Realism

  • What follows are some premises for data science magical realism stories based (very, very loosely) on experiences I’ve had or heard about — premises, that is, for stories about impossible, absurd, magical things happening to data scientists in ordinary data science situations. Enjoy!
  • More from David Masad

Program Synthesis in 2017-18

  • A high-level overview of the recent ideas and representative papers in program synthesis as of mid-2018.
  • Alex (Oleksandr) Polozov, a researcher in the Deep Procedural Intelligence group at Microsoft Research AI, Redmond. I work on neural program synthesis from input-output examples and natural language, intersections of machine learning and software engineering, and neuro-symbolic architectures. I am particularly interested in combining neural and symbolic techniques to tackle the next generation of AI problems, including program synthesis, planning, and reasoning.

UMAP Uniform Manifold Approximation and Projection for Dimension Reduction | SciPy 2018 |(video) (paper)

  • UMAP (Uniform Manifold Approximation and Projection) is a novel manifold learning technique for dimension reduction. UMAP is constructed from a theoretical framework based in Riemannian geometry and algebraic topology. The result is a practical scalable algorithm that applies to real world data. The UMAP algorithm is competitive with t-SNE for visualization quality, and arguably preserves more of the global structure with superior run time performance. Furthermore, UMAP as described has no computational restrictions on embedding dimension, making it viable as a general purpose dimension reduction technique for machine learning.
  • This could be nice for building maps

7:00 – 5:00 ASRC MKT

  • Progress on getting my keys back!
  • Got everyone’s response on the Doodle, but only 4 of the 5 line up…
  • Finish first pass through PhD review slides
  • Start SASO slides and poster?
  • Continue with exporting terms from the sim and importing them into python. One of the things that will matter is the tagging of the data with the seed terms from the sim as well as the cell name so that reconstructions can be compared for accuracy.
  • Added the cell location to each <sampleData> so that there can be some kind of tagging/ground truth about the maps we’re inferring.
  • Working on iterating through the etree hierarchy. I can now read in the file, parse it and get elements that I’m looking for.
  • Tomorrow will be pulling the seed words out of the code in an ordered list. Generated sentences will need to be timestamped to that conversations can be reconstructed. That being said, it could be interesting to take seed words out of a generated sentence and add them to the embedding seed words. Something to think about.

Phil 6.22.18

7:00 – 5:30 ASRC MKT

  • Twitter experiment on a fake Gary Indiana secession. IFTTT retweeting leads to interesting behavior.
  • Fixed FlockingShape casting by adding a customDrawStep(GraphicsContext gc) to the SmartShape base class that’s called from draw().
  • Add records to each agent that store a list of source and agent influences at each time sample. It should include the name of the item and the amount of influence. Probably save as an XML file, since it has too many dimensions. The file could then be used to create terms or spreadsheets.
    • Started on CAInfluence class which will be added to CA classes in an arrayList in BaseCA;
  • More file conversion with Bob – and everything worked great until I try one after Bob leaves. Ka-BOOM!
    • Installed all the packages to get everything to run in the debugger. Found what appears to be a perfectly good line “range” that causes the problem? Will start debugging on Wednesday.
  • Project MERCATOR proposal
  • Meeting with Sy

Phil 6.21.18

7:00 – 4:00 ASRC MKT

  • Add an attractor scalar for agents that’s normally zero. A vector to each agent within the SIH is calculated and scaled by the attractor scalar. That vector is then added to the direction vector to the agent – done
  • Remove the heading influence based on site – done
  • Add a white circle to the center of the agent that is the size of the attraction scalar. Done
  • Add attraction radius slider that is independent of the SIH. -done
  • Add a ‘site trajectory’ to the spreadsheet that will have the site lists (and their percentage?)
  • There is now an opportunity for a poster and a demo at SASO
  • Add stories, lists and maps to implication slides – done
  • Got all my connections set up
  • Successfully converted and deployed cosmos-2
  • Voted!

Phil 6.20.18

7:00 – 9:00 2:00 – 5:00 ASRC MKT

  • Redo doodle for all of August – done
  • Schooling Fish May Offer Insights Into Networked Neurons
    • Iain Couzin is deciphering the rules that govern group behavior. The results might provide a fresh perspective on how networks of neurons work together.
  • City arts and lectures: The New Science Of Psychedelics With Michael Pollan
    • Psychedelics reduce the section of the brain that have to do with the sense of self. Pollan thinks that this also happens with certain types of rhythmic music and in crowd situations. This could be related to stampedes and flocking.
    • LSD May Chip Away at the Brain’s “Sense of Self” Network
      • Brain imaging suggests LSD’s consciousness-altering traits may work by hindering some brain networks and boosting overall connectivity
  • Add an attractor scalar for agents that’s normally zero. A vector to each agent within the SIH is calculated and scaled by the attractor scalar. That vector is then added to the direction vector to the agent – done?
  • Remove the heading influence based on site – done
  • Add a white circle to the center of the agent that is the size of the attraction scalar. Done
  • Add a ‘site trajectory’ to the spreadsheet that will have the site lists (and their percentage?)
  • Worked on A2P white paper with Aaron.
  • Worked on a response to Dr. Li’s response

ASRC IRAD 9:00 – 2:00

  • Mind meld with Bob
    • Revisit Yarn
    • Excel stuff?
    • Connect to AWS using bastion. Look in FoxyProxy how to. I need certs
    • Drop on rabbit to deploy to CI and QA and NESDIS  ONE (production)
    • Don’t want sensitive information in Git. We use sharepoint instead
    • Notes and screenshots in document.

Phil 6.11.18

7:00 – 6:00 ASRC MKT

  • More Bit by Bit. Reading the section on ethics. It strikes me that simulation could be a way to cut the PII Gordion Knot in some conditions. If a simulation can be developed that generates statistically similar data to the desired population, then the simulated data and the simulation code can be released to the research community. The dataset becomes infinite and adjustable, while the PII data can be held back. Machine learning systems trained on the simulated data can then be evaluated on the confidential data. The differences in the classification by the ML systems between real data and simulated data can also provide insight into the gaps in fidelity of the simulated data, which would provide an ongoing improvement to the simulation, which could in turn be released to the community.
  • Continuing with the cleanup of the SASO paper. Mostly done but some trimming of redundent bits and the “Ose Simple Trick” paragraph.
  • SASO travel link
    • Monday prices: SASO
  • Fika
    • Come up with 3-5 options for a finished state for the dissertation. It probably ranges from “pure theory” through “instance based on theory” to “a map generated by the system that matches the theory”
    • Once the SASO paper is in, set up a “wine and cheese” get together for the committee to go over the current work and discuss changes to the next phase
    • Start on a new IRB. Emphasize how everyone will have the same system to interact with, though their interactions will be different. Emphasize that the system has to allow open interaction to provide the best chance to realize theoretical results.
    • Will and I are on the hook for a Fika about LaTex

Phil 5.15.18

7:00 – 4:00 ASRC MKT