Category Archives: Python

Phil 9.17.19

7:00 – 6:00 ASRC GOES

  • Dept of vital records 410 764 2922 maryland.gov
  • Working from home today, waiting for a delivery
  • Meet with Will at 3:00. Went smoothly this time.
  • Send Aaron a note that I’ll miss next week and maybe the week after
  • Dissertation – slow going. Wrote a few paragraphs on lists and stories. Need to put the section together on games
  • EvolutionaryOptimizer – done? It’s working nicely on the test set. You can see that it doesn’t always come up with the best answer, but it’s always close and often much faster:
  • Need to write the fitness function that builds and evaluates the model
  • Worked on getting TF 2.0 installed using my instructions, but the TF 2.0 build is broken? Ah, I see that we are now at RC1. Changing the instructions.
  • Everything works now, but my day is done. Need to update my install at work tomorrow.

 

Phil 9.16.19

7:00 – 8:00 ASRC GOES

This makes me happy. Older, but not slower. Yet.

Strave

  • Maryland Anatomy Board Dept of vital records 410 764 2922 – Never got called back
  • Ping Antonio about virtual crowdsourcing of opinion
  • Dissertation – write up dissertation house one-pager
  • Optimizer
    • Generating chromosome sequences.
    • Created a fitness landscape to evaluate

FitnessLandscape

  • Working on breeding and mutation
  • ML Seminar
    • Status, and a few more Andrew Ng segments. How to debug gradient descent
  • Meeting With Aaron M
    • Nice chat
    • GARY MARCUS is a scientist, best-selling author, and entrepreneur. He is Founder and CEO of Robust.AI.
    • His newest book, co-authored with Ernest Davis, Rebooting AI: Building Machines We Can Trust aims to shake up the field of artificial intelligence.
    • Don’t put the transformer research in the dissertation
  • Evolution of Representations in the Transformer (nice looking blog post of deeper paper)
    • We look at the evolution of representations of individual tokens in Transformers trained with different training objectives (MT, LM, MLM – BERT-style) from the Information Bottleneck perspective and show, that:
      • LMs gradually forget past when forming future;
      • for MLMs, the evolution has the two stages of context encoding and token reconstruction;
      • MT representations get refined with context, but less processing is happening.
  • Different Spirals of Sameness: A Study of Content Sharing in Mainstream and Alternative Media
    • In this paper, we analyze content sharing between news sources in the alternative and mainstream media using a dataset of 713K articles and 194 sources. We find that content sharing happens in tightly formed communities, and these communities represent relatively homogeneous portions of the media landscape. Through a mix-method analysis, we find several primary content sharing behaviors. First, we find that the vast majority of shared articles are only shared with similar news sources (i.e. same community). Second, we find that despite these echo-chambers of sharing, specific sources, such as The Drudge Report, mix content from both mainstream and conspiracy communities. Third, we show that while these differing communities do not always share news articles, they do report on the same events, but often with competing and counter-narratives. Overall, we find that the news is homogeneous within communities and diverse in between, creating different spirals of sameness.

Phil 9.12.19

7:00 – 4:30 ASRC GOES

  • FractalNet: Ultra-Deep Neural Networks without Residuals
    • We introduce a design strategy for neural network macro-architecture based on self-similarity. Repeated application of a simple expansion rule generates deep networks whose structural layouts are precisely truncated fractals. These networks contain interacting subpaths of different lengths, but do not include any pass-through or residual connections; every internal signal is transformed by a filter and nonlinearity before being seen by subsequent layers. In experiments, fractal networks match the excellent performance of standard residual networks on both CIFAR and ImageNet classification tasks, thereby demonstrating that residual representations may not be fundamental to the success of extremely deep convolutional neural networks. Rather, the key may be the ability to transition, during training, from effectively shallow to deep. We note similarities with student-teacher behavior and develop drop-path, a natural extension of dropout, to regularize co-adaptation of subpaths in fractal architectures. Such regularization allows extraction of high-performance fixed-depth subnetworks. Additionally, fractal networks exhibit an anytime property: shallow subnetworks provide a quick answer, while deeper subnetworks, with higher latency, provide a more accurate answer.
  • Structural diversity in social contagion
    • The concept of contagion has steadily expanded from its original grounding in epidemic disease to describe a vast array of processes that spread across networks, notably social phenomena such as fads, political opinions, the adoption of new technologies, and financial decisions. Traditional models of social contagion have been based on physical analogies with biological contagion, in which the probability that an individual is affected by the contagion grows monotonically with the size of his or her “contact neighborhood”—the number of affected individuals with whom he or she is in contact. Whereas this contact neighborhood hypothesis has formed the underpinning of essentially all current models, it has been challenging to evaluate it due to the difficulty in obtaining detailed data on individual network neighborhoods during the course of a large-scale contagion process. Here we study this question by analyzing the growth of Facebook, a rare example of a social process with genuinely global adoption. We find that the probability of contagion is tightly controlled by the number of connected components in an individual’s contact neighborhood, rather than by the actual size of the neighborhood. Surprisingly, once this “structural diversity” is controlled for, the size of the contact neighborhood is in fact generally a negative predictor of contagion. More broadly, our analysis shows how data at the size and resolution of the Facebook network make possible the identification of subtle structural signals that go undetected at smaller scales yet hold pivotal predictive roles for the outcomes of social processes.
    • Add this to the discussion section – done
  • Dissertation
    • Started on the theory section, then realized the background section didn’t set it up well. So worked on the background instead. I put in a good deal on how individuals and groups interact with the environment differently and how social interaction amplifies individual contribution through networking.
  • Quick meetings with Don and Aaron
  • Time prediction (sequence to sequence) with Keras perceptrons
  • This was surprisingly straightforward
    • There was some initial trickiness in getting the IDE to work with the TF2.0 RC0 package:
      import tensorflow as tf
      from tensorflow import keras
      from tensorflow_core.python.keras import layers

      The first coding step was to generate the data. In this case I’m building a numpy matrix that has ten variations on math.sin(), using our timeseriesML utils code. There is a loop that sets up the code to create a new frequency, which is sent off to get back a pandas Dataframe that in this case has 10 sequence rows with 100 samples. First, we set the global sequence_length:

      sequence_length = 100

      then we create the function that will build and concatenate our numpy matrices:

      def generate_train_test(num_functions, rows_per_function, noise=0.1) -> (np.ndarray, np.ndarray, np.ndarray):
          ff = FF.float_functions(rows_per_function, 2*sequence_length)
          npa = None
          for i in range(num_functions):
              mathstr = "math.sin(xx*{})".format(0.005*(i+1))
              #mathstr = "math.sin(xx)"
              df2 = ff.generateDataFrame(mathstr, noise=0.1)
              npa2 = df2.to_numpy()
              if npa is None:
                  npa = npa2
              else:
                  ta = np.append(npa, npa2, axis=0)
                  npa = ta
      
          split = np.hsplit(npa, 2)
          return npa, split[0], split[1]

      Now, we build the model. We’re using keras from the TF 2.0 RC0 build, so things look slightly different:

      model = tf.keras.Sequential()
      # Adds a densely-connected layer with 64 units to the model:
      model.add(layers.Dense(sequence_length, activation='relu', input_shape=(sequence_length,)))
      # Add another:
      model.add(layers.Dense(200, activation='relu'))
      # Add a softmax layer with 10 output units:
      model.add(layers.Dense(sequence_length))
      
      loss_func = tf.keras.losses.MeanSquaredError()
      opt_func = tf.keras.optimizers.Adam(0.01)
      model.compile(optimizer= opt_func,
                    loss=loss_func,
                    metrics=['accuracy'])

      We can now fit the model to the generated data:

      full_mat, train_mat, test_mat = generate_train_test(10, 10)
      
      model.fit(train_mat, test_mat, epochs=10, batch_size=2)

      There is noise in the data, so the accuracy is not bang on, but the loss is nice. We can see this better in the plots above, which were created using this function:

      def plot_mats(mat:np.ndarray, cluster_size:int, title:str, fig_num:int):
          plt.figure(fig_num)
      
          i = 0
          for row in mat:
              cstr = "C{}".format(int(i/cluster_size))
              plt.plot(row, color=cstr)
              i += 1
      
          plt.title(title)

      Which is called just before the program completes:

      if show_plots:
          plot_mats(full_mat, 10, "Full Data", 1)
          plot_mats(train_mat, 10, "Input Vector", 2)
          plot_mats(test_mat, 10, "Output Vector", 3)
          plot_mats(predict_mat, 10, "Predict", 4)
          plt.show()
    • That’s it! Full listing below:
import tensorflow as tf
from tensorflow import keras
from tensorflow_core.python.keras import layers
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import timeseriesML.generators.float_functions as FF


sequence_length = 100

def generate_train_test(num_functions, rows_per_function, noise=0.1) -> (np.ndarray, np.ndarray, np.ndarray):
    ff = FF.float_functions(rows_per_function, 2*sequence_length)
    npa = None
    for i in range(num_functions):
        mathstr = "math.sin(xx*{})".format(0.005*(i+1))
        #mathstr = "math.sin(xx)"
        df2 = ff.generateDataFrame(mathstr, noise=0.1)
        npa2 = df2.to_numpy()
        if npa is None:
            npa = npa2
        else:
            ta = np.append(npa, npa2, axis=0)
            npa = ta

    split = np.hsplit(npa, 2)
    return npa, split[0], split[1]

def plot_mats(mat:np.ndarray, cluster_size:int, title:str, fig_num:int):
    plt.figure(fig_num)

    i = 0
    for row in mat:
        cstr = "C{}".format(int(i/cluster_size))
        plt.plot(row, color=cstr)
        i += 1

    plt.title(title)

model = tf.keras.Sequential()
# Adds a densely-connected layer with 64 units to the model:
model.add(layers.Dense(sequence_length, activation='relu', input_shape=(sequence_length,)))
# Add another:
model.add(layers.Dense(200, activation='relu'))
# Add a softmax layer with 10 output units:
model.add(layers.Dense(sequence_length))

loss_func = tf.keras.losses.MeanSquaredError()
opt_func = tf.keras.optimizers.Adam(0.01)
model.compile(optimizer= opt_func,
              loss=loss_func,
              metrics=['accuracy'])

full_mat, train_mat, test_mat = generate_train_test(10, 10)

model.fit(train_mat, test_mat, epochs=10, batch_size=2)
model.evaluate(train_mat, test_mat)

# test against freshly generated data
full_mat, train_mat, test_mat = generate_train_test(10, 10)
predict_mat = model.predict(train_mat)

show_plots = True
if show_plots:
    plot_mats(full_mat, 10, "Full Data", 1)
    plot_mats(train_mat, 10, "Input Vector", 2)
    plot_mats(test_mat, 10, "Output Vector", 3)
    plot_mats(predict_mat, 10, "Predict", 4)
    plt.show()



Phil 9.11 . 19

be7c6582-044a-4a19-aa8b-de388b4a4f83-cincpt_09-11-2016_enquirer_1_b001__2016_09_10_img_xxx_world_trade_11_1_1_9kfm0g4g_l880019336_img_xxx_world_trade_11_1_1_9kfm0g4g

7:00 – 4:00 ASRC GOES

  • Model:DLG3501W SKU:6181264
  • Maryland Anatomy Board Dept of vital records 410 764 2922
  • arxiv-vanity.com  arXiv Vanity renders academic papers from arXiv as responsive web pages so you don’t have to squint at a PDF.
    • It works ok. Tables and cation alignment are a problem for now, but it sounds great for phones
  • DeepPrivacy: A Generative Adversarial Network for Face Anonymization
    • We propose a novel architecture which is able to automatically anonymize faces in images while retaining the original data distribution. We ensure total anonymization of all faces in an image by generating images exclusively on privacy-safe information. Our model is based on a conditional generative adversarial network, generating images considering the original pose and image background. The conditional information enables us to generate highly realistic faces with a seamless transition between the generated face and the existing background. Furthermore, we introduce a diverse dataset of human faces, including unconventional poses, occluded faces, and a vast variability in backgrounds. Finally, we present experimental results reflecting the capability of our model to anonymize images while preserving the data distribution, making the data suitable for further training of deep learning models. As far as we know, no other solution has been proposed that guarantees the anonymization of faces while generating realistic images.
  • Introducing a Conditional Transformer Language Model for Controllable Generation
    • CTRL is a 1.6 billion-parameter language model with powerful and controllable artificial text generation that can predict which subset of the training data most influenced a generated text sequence. It provides a potential method for analyzing large amounts of generated text by identifying the most influential source of training data in the model. Trained with over 50 different control codes, the CTRL model allows for better human-AI interaction because users can control the generated content and style of the text, as well as train it for multitask language generation. Finally, it can be used to improve other natural language processing (NLP) applications either through fine-tuning for a specific task or through transfer of representations that the model has learned.
  • Dissertation
    • Started to put together my Linux laptop for vacation writing
    • More SIH section
  • Verify that timeseriesML can be used as a library
  • Perceptron curve prediction
  • AI/ML status meetings
  • Helped Vadim with some python issues

Phil 9.10.19

ASRC GOES 7:00 – 5:30

  • Got a mention in an article on Albawaba – When the Only Option is ‘Not to Play’? Autonomous Weapons Systems Debated in Geneva 
  • Dissertation – more SIH
  • Just saw this: On Extractive and Abstractive Neural Document Summarization with Transformer Language Models
    • We present a method to produce abstractive summaries of long documents that exceed several thousand words via neural abstractive summarization. We perform a simple extractive step before generating a summary, which is then used to condition the transformer language model on relevant information before being tasked with generating a summary. We show that this extractive step significantly improves summarization results. We also show that this approach produces more abstractive summaries compared to prior work that employs a copy mechanism while still achieving higher rouge scores. Note: The abstract above was not written by the authors, it was generated by one of the models presented in this paper.
  • Working on packaging timeseriesML. I think it’s working!

TimeSeriesML

  • I’ll try it out when I get back after lunch
  • Meeting with Vadim
    • Showed him around and provided svn access
  • Model:DLG3501W SKU:6181264

Phil 9.5.19

7:00 –

  • David Manheim (scholar)
    • I work on existential risk mitigation, computational modelling, and epidemiology. I spend time talking about Goodhart’s Law, and have been a #Superforecaster with the Good Judgement Project since 2012.
  • Goodhart’s law is an adage named after economist Charles Goodhart, which has been phrased by Marilyn Strathern as “When a measure becomes a target, it ceases to be a good measure.”[1] One way in which this can occur is individuals trying to anticipate the effect of a policy and then taking actions that alter its outcome
  • Dissertation
  • Continuing TF 2.0 Keras tutorial
    • Had a weird problem where
      from tensorflow import keras

      made IntelliJ complain, but the python interpreter ran fine. I then installed keras, and IJ stopped complaining. Checking the version(s) seems to be identical, even though I can see that there is a new keras directory in D:\Program Files\Python37\Lib\site-packages. And we know that the interpreter and IDE are pointing to the same place:

      "D:\Program Files\Python37\python.exe" D:/Development/Sandboxes/PyBullet/src/TensorFlow/HelloKeras.py
      2019-09-05 11:30:04.694327: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library cudart64_100.dll
      tf.keras version = 2.2.4-tf
      keras version = 2.2.4-tf

Keras

    • This has the implication that instead of :
      from tensorflow.keras import layers

      I need to have:

      from keras import layers

      I mean, it works, but it’s weird and makes me think that something subtle may be busted…

Phil 9.4.19

7:00 – 5:00 ASRC GOES

Surrogation

Phil 8.30.19

7:00 – 4:00 ASRC GOES

  • Dentist!
  • Sent notes to David Lazar and Erika M-T. Still need to ping Stuart Shulman.
  • Did my part for JuryRoom (Eero Mäntyranta)
  • Dissertation – more on State
  • TF 2.0 today? (release notes)
  • Installed! Well – it didn’t blow up…
    C:\WINDOWS\system32>pip3 install tensorflow-gpu==2.0.0-rc0                                                              Collecting tensorflow-gpu==2.0.0-rc0                                                                                      Downloading https://files.pythonhosted.org/packages/3c/90/046fdf56ba957de792e4132b687e09e34b6f237608aa9fc17c656ab69b39/tensorflow_gpu-2.0.0rc0-cp37-cp37m-win_amd64.whl (285.1MB)                                                                  |████████████████████████████████| 285.1MB 20kB/s                                                                  Collecting absl-py>=0.7.0 (from tensorflow-gpu==2.0.0-rc0)                                                                Downloading https://files.pythonhosted.org/packages/3c/0d/7cbf64cac3f93617a2b6b079c0182e4a83a3e7a8964d3b0cc3d9758ba002/absl-py-0.8.0.tar.gz (102kB)                                                                                                |████████████████████████████████| 112kB ...                                                                       Collecting gast>=0.2.0 (from tensorflow-gpu==2.0.0-rc0)                                                                   Downloading https://files.pythonhosted.org/packages/4e/35/11749bf99b2d4e3cceb4d55ca22590b0d7c2c62b9de38ac4a4a7f4687421/gast-0.2.2.tar.gz                                                                                                      Collecting google-pasta>=0.1.6 (from tensorflow-gpu==2.0.0-rc0)                                                           Downloading https://files.pythonhosted.org/packages/d0/33/376510eb8d6246f3c30545f416b2263eee461e40940c2a4413c711bdf62d/google_pasta-0.1.7-py3-none-any.whl (52kB)                                                                                  |████████████████████████████████| 61kB 4.1MB/s                                                                    Collecting wrapt>=1.11.1 (from tensorflow-gpu==2.0.0-rc0)                                                                 Downloading https://files.pythonhosted.org/packages/23/84/323c2415280bc4fc880ac5050dddfb3c8062c2552b34c2e512eb4aa68f79/wrapt-1.11.2.tar.gz                                                                                                    Collecting grpcio>=1.8.6 (from tensorflow-gpu==2.0.0-rc0)                                                                 Downloading https://files.pythonhosted.org/packages/32/e7/478737fd426798caad32a2abb7cc63ddb4c12908d9e03471dd3c41992b05/grpcio-1.23.0-cp37-cp37m-win_amd64.whl (1.6MB)                                                                              |████████████████████████████████| 1.6MB ...                                                                       Collecting termcolor>=1.1.0 (from tensorflow-gpu==2.0.0-rc0)                                                              Downloading https://files.pythonhosted.org/packages/8a/48/a76be51647d0eb9f10e2a4511bf3ffb8cc1e6b14e9e4fab46173aa79f981/termcolor-1.1.0.tar.gz                                                                                                 Collecting tf-estimator-nightly<1.14.0.dev2019080602,>=1.14.0.dev2019080601 (from tensorflow-gpu==2.0.0-rc0)              Downloading https://files.pythonhosted.org/packages/21/28/f2a27a62943d5f041e4a6fd404b2d21cb7c59b2242a4e73b03d9ba166552/tf_estimator_nightly-1.14.0.dev2019080601-py2.py3-none-any.whl (501kB)                                                      |████████████████████████████████| 501kB ...                                                                       Collecting keras-preprocessing>=1.0.5 (from tensorflow-gpu==2.0.0-rc0)                                                    Downloading https://files.pythonhosted.org/packages/28/6a/8c1f62c37212d9fc441a7e26736df51ce6f0e38455816445471f10da4f0a/Keras_Preprocessing-1.1.0-py2.py3-none-any.whl (41kB)                                                                       |████████████████████████████████| 51kB 3.2MB/s                                                                    Collecting tb-nightly<1.15.0a20190807,>=1.15.0a20190806 (from tensorflow-gpu==2.0.0-rc0)                                  Downloading https://files.pythonhosted.org/packages/bc/88/24b5fb7280e74c7cf65bde47c171547fd02afb3840cff41bcbe9270650f5/tb_nightly-1.15.0a20190806-py3-none-any.whl (4.3MB)                                                                         |████████████████████████████████| 4.3MB 6.4MB/s                                                                   Collecting wheel>=0.26 (from tensorflow-gpu==2.0.0-rc0)                                                                   Downloading https://files.pythonhosted.org/packages/00/83/b4a77d044e78ad1a45610eb88f745be2fd2c6d658f9798a15e384b7d57c9/wheel-0.33.6-py2.py3-none-any.whl                                                                                      Requirement already satisfied: numpy<2.0,>=1.16.0 in d:\program files\python37\lib\site-packages (from tensorflow-gpu==2.0.0-rc0) (1.16.4)                                                                                                      Collecting protobuf>=3.6.1 (from tensorflow-gpu==2.0.0-rc0)                                                               Downloading https://files.pythonhosted.org/packages/46/8b/5e77963dac4a944a0c6b198c004fac4c85d7adc54221c288fc6ca9078072/protobuf-3.9.1-cp37-cp37m-win_amd64.whl (1.0MB)                                                                             |████████████████████████████████| 1.0MB 6.4MB/s                                                                   Requirement already satisfied: six>=1.10.0 in d:\program files\python37\lib\site-packages (from tensorflow-gpu==2.0.0-rc0) (1.12.0)                                                                                                             Collecting opt-einsum>=2.3.2 (from tensorflow-gpu==2.0.0-rc0)                                                             Downloading https://files.pythonhosted.org/packages/c0/1a/ab5683d8e450e380052d3a3e77bb2c9dffa878058f583587c3875041fb63/opt_einsum-3.0.1.tar.gz (66kB)                                                                                              |████████████████████████████████| 71kB 4.5MB/s                                                                    Collecting astor>=0.6.0 (from tensorflow-gpu==2.0.0-rc0)                                                                  Downloading https://files.pythonhosted.org/packages/d1/4f/950dfae467b384fc96bc6469de25d832534f6b4441033c39f914efd13418/astor-0.8.0-py2.py3-none-any.whl                                                                                       Collecting keras-applications>=1.0.8 (from tensorflow-gpu==2.0.0-rc0)                                                     Downloading https://files.pythonhosted.org/packages/71/e3/19762fdfc62877ae9102edf6342d71b28fbfd9dea3d2f96a882ce099b03f/Keras_Applications-1.0.8-py3-none-any.whl (50kB)                                                                            |████████████████████████████████| 51kB 3.2MB/s                                                                    Collecting markdown>=2.6.8 (from tb-nightly<1.15.0a20190807,>=1.15.0a20190806->tensorflow-gpu==2.0.0-rc0)                 Downloading https://files.pythonhosted.org/packages/c0/4e/fd492e91abdc2d2fcb70ef453064d980688762079397f779758e055f6575/Markdown-3.1.1-py2.py3-none-any.whl (87kB)                                                                                  |████████████████████████████████| 92kB 5.8MB/s                                                                    Collecting setuptools>=41.0.0 (from tb-nightly<1.15.0a20190807,>=1.15.0a20190806->tensorflow-gpu==2.0.0-rc0)              Downloading https://files.pythonhosted.org/packages/b2/86/095d2f7829badc207c893dd4ac767e871f6cd547145df797ea26baea4e2e/setuptools-41.2.0-py2.py3-none-any.whl (576kB)                                                                              |████████████████████████████████| 583kB ...                                                                       Collecting werkzeug>=0.11.15 (from tb-nightly<1.15.0a20190807,>=1.15.0a20190806->tensorflow-gpu==2.0.0-rc0)               Downloading https://files.pythonhosted.org/packages/d1/ab/d3bed6b92042622d24decc7aadc8877badf18aeca1571045840ad4956d3f/Werkzeug-0.15.5-py2.py3-none-any.whl (328kB)                                                                                |████████████████████████████████| 337kB 6.4MB/s                                                                   Collecting h5py (from keras-applications>=1.0.8->tensorflow-gpu==2.0.0-rc0)                                               Downloading https://files.pythonhosted.org/packages/4f/1e/89aa610afce8df6fd1f12647600a05e902238587ae6375442a3164b59d51/h5py-2.9.0-cp37-cp37m-win_amd64.whl (2.4MB)                                                                                 |████████████████████████████████| 2.4MB ...                                                                       Installing collected packages: absl-py, gast, google-pasta, wrapt, grpcio, termcolor, tf-estimator-nightly, keras-preprocessing, setuptools, markdown, werkzeug, wheel, protobuf, tb-nightly, opt-einsum, astor, h5py, keras-applications, tensorflow-gpu                                                                                                                 Running setup.py install for absl-py ... done                                                                           Running setup.py install for gast ... done                                                                              Running setup.py install for wrapt ... done                                                                             Running setup.py install for termcolor ... done                                                                         Found existing installation: setuptools 40.8.0                                                                            Uninstalling setuptools-40.8.0:                                                                                           Successfully uninstalled setuptools-40.8.0                                                                          Running setup.py install for opt-einsum ... done                                                                      Successfully installed absl-py-0.8.0 astor-0.8.0 gast-0.2.2 google-pasta-0.1.7 grpcio-1.23.0 h5py-2.9.0 keras-applications-1.0.8 keras-preprocessing-1.1.0 markdown-3.1.1 opt-einsum-3.0.1 protobuf-3.9.1 setuptools-41.2.0 tb-nightly-1.15.0a20190806 tensorflow-gpu-2.0.0rc0 termcolor-1.1.0 tf-estimator-nightly-1.14.0.dev2019080601 werkzeug-0.15.5 wheel-0.33.6 wrapt-1.11.2

     

  • Oops: Python 3.7.4 (tags/v3.7.4:e09359112e, Jul 8 2019, 20:34:20) [MSC v.1916 64 bit (AMD64)] on win32
    Type “help”, “copyright”, “credits” or “license” for more information.
    >>> import tensorflow as tf
    2019-08-30 10:23:30.632254: W tensorflow/stream_executor/platform/default/dso_loader.cc:55] Could not load dynamic library ‘cudart64_100.dll’; dlerror: cudart64_100.dll not found
    >>>
  • Spent the rest of the day trying to get “import tensorflow as tf” to work

Phil 8.29.19

ASRC GOES – 7:00 – 4:00

  • Find out who I was talking to yesterday at lunch (Boynton?)
  • Contact David Lazar about RB
  • Participating as an anonymous fish in JuryRoom. Started the discussion
  • Dissertation – started the State section
  • Working on Control and sim diagrams
    • Putting this here because I keep on forgetting how to add an outline/border to an image in Illustrator:

OutlineAI

  1. Place and select an image in the Illustrator document.
  2. Once selected, open Appearance panel and from the Appearance panel flyout menu, choose Add New Stroke:
  3. With the Stroke highlighted in the Appearance panel, choose Effect -> Path -> Outline Object.
  • Anyway, back to our regularly scheduled program.
  • Made a control system diagram
  • Made a control system inheritance diagram
  • Made a graphics inheritance diagram
  • Need to stick them in the ASRC Dev Pipeline document
  • Discovered JabRef: JabRef is an open source bibliography reference manager. The native file format used by JabRef is BibTeX, the standard LaTeX bibliography format. JabRef is a desktop application and runs on the Java VM (version 8), and works equally well on Windows, Linux, and Mac OS X.
  • Tomorrow we get started with TF 2.0

Phil 9.27.19

7:00 – 7:00 ASRC GOES

  • Replied to Antonio with a plan for the software paper
  • Dissertation
    • More Lit Review – finished Dimension Reduction!
  • Set up TF 2.0 – Nope
  • Started writeup of simulator and sent a status to Erik
  • Waikato meeting
    • First room is running
    • Asked Chris to change “Maximum Anonymity” to “Improved Anonymity”
    • Some discussion about experimental design

Phil 8.26.19

7:00 – ASRC GOES

  • Dissertation – working my way through the lit review section
  • Antonio sent a note about Software Impacts, which provides a scholarly reference to software that has been used to address a research challenge. The journal disseminates impactful and re-usable scientific software through Original Software Publications (OSP) which describe the application of the software to research and the published outputs.
    • Submissions to Software Impacts consist of two major parts:
      • A short descriptive paper of about three pages including an Impact Overview and references to publications where the software has been used
      • An open source software distribution with support material.
    • So, to get things to fit on GitHub, I worked on getting GPM to work with a smaller library – done
  • Discussions with Aaron about using TF 2.0 xformer on GOES sim data
  • Security training – an hour or so
  • Copied the Waikato JuryRoom proposal to PolarizationGame folder

Phil 8.23.19

7:00 – 4:00 ASRC GEOS

  • More Dissertation
    • Continuing lit review
  • Rework BlueSky paper for air traffic? Meeting with T at 10:00
  • Simulation
    • Need to discuss with Aaron the best way to use the data to train the NN and round-trip the outputs so that they can be used to have the ML model issue commands to the RCS system so that given the outputs of one model, the NN can create commands that cause the same outputs in a separate model
  • Wow. It knows/finds syntactically correct Java. From TalkToTransformer.com:
  • Wow

Phil 8.22.19

7:00 – ASRC GOES

ScottW

  • Dissertation
    • Lit review
    • This, from Colin Martindale CogPsy a NN approach. It’s the central piece:
      • it turns out that language is almost entirely metaphorical (Hobbs,
        1983; Lakoff, 1987; Lakoff & Johnson, 1980). Many of these metaphors are
        spatial. Look back at the last sentence. I asked you to think things through. I told you that something turned out. We bring up topics. We put them on the table. If you could argue with me, Lakoff and Johnson (1980) point out that we would have a war: you might try to attack and shoot down my arguments. I would try to defend them by trying to demolish your position and counterattacking. Lakoff’s argument is that if we took all the metaphors out of language, there would be virtually nothing left. (p 212)
  • More control systems – first pass is working!

RunningSim

InputVector