Greg Yang

@TheGregYang

Researcher at AI. Morgan Prize Honorable Mention 2018.

Bellevue, WA
Vrijeme pridruživanja: veljača 2019.

Tweetovi

Blokirali ste korisnika/cu @TheGregYang

Jeste li sigurni da želite vidjeti te tweetove? Time nećete deblokirati korisnika/cu @TheGregYang

  1. Prikvačeni tweet
    5. pro 2019.

    1/ Why do wide, random neural networks form Gaussian processes, *regardless of architecture*? Let me give an overview in case you are too lazy to check out the paper or the code . The proof has two parts…

    Prikaži ovu nit
    Poništi
  2. proslijedio/la je Tweet

    With projects ranging in focus from healthcare to gaming to server workload management, 2019-20 Microsoft AI residents are pushing forward real-world applications in artificial intelligence. Learn more about the program and apply for the class of 2020-21:

    Poništi
  3. proslijedio/la je Tweet

    Dr. leads the Deep Learning Group at Microsoft Research. Hear how recent advances in training data and infrastructure have allowed Dr. Gao’s team to work on applications for deep learning in language, vision, and video on the :

    Poništi
  4. proslijedio/la je Tweet
    28. sij

    This is a great question that I've gotten periodically. Previously it would have taken too long to put something together, but using Neural Tangents () it's really easy and fast! Here is the reproduction in a colab:

    Prikaži ovu nit
    Poništi
  5. proslijedio/la je Tweet

    With TensorWatch, researchers seek to help other researchers/engineers get much-needed info about the state of their ML systems. The open-source debugging and visualization tool works in Jupyter Notebook to perform key analysis tasks for models and data:

    Poništi
  6. proslijedio/la je Tweet
    24. sij

    Research on the Neural Tangent Kernel (NTK) almost exclusively uses a non-standard neural network parameterization, where activations are divided by sqrt(width), and weights are initialized to have variance 1 rather than variance 1/width.

    Prikaži ovu nit
    Poništi
  7. proslijedio/la je Tweet
    23. sij

    Super disgusted by people working on privacy-related technologies who knowingly misrepresent "their" technology to the employer or policymakers hoping for a faster or wider adoption. Such people make matters strictly worse for everyone (except possibly the company they work for).

    Poništi
  8. 21. sij

    Joint work with these awesome algebraicists and geometers Justin Chen, Christopher Eur, and, Mengyuan Zhang!

    Prikaži ovu nit
    Poništi
  9. 21. sij
    Prikaži ovu nit
    Poništi
  10. 21. sij

    Learnability (VC dim) is a *topological property*, as I proved in for parity, conjunctions, poly threshold fctns. Now this extends to downward-closed classes, conjunction of parities, and k-CNFs, as well! Just how far does this go?

    Prikaži ovu nit
    Poništi
  11. proslijedio/la je Tweet

    Text-based games provide a platform to train RL agents that generate goal-driven language. Jericho framework by & provides benchmarks for scaling RL to combinatorially sized language action spaces:

    Poništi
  12. 13. sij

    Did u know? The NTK & GP kernels of an MLP on the sphere have the form K(x, y) = F(x.y/dim) for scalar F. Each (eigenvalue*multiplicity) of K goes to F^{(k)}(0)/k! as dim -> infty, over unif dist on the sphere (or the bool cube, also extends to Gaussian)

    Poništi
  13. proslijedio/la je Tweet

    Microsoft AirSim, now on Unity, provides an open-source system for training autonomous systems. Faster and safer than training these systems in the real world, explore how AirSim utilizes multiple learning methods to create a realistic environment:

    Poništi
  14. proslijedio/la je Tweet
    8. sij

    A few weeks ago, published some interesting research on how smiles can serve as feedback for machine learning. I don't think it got enough attention.

    Poništi
  15. 7. sij

    Neural networks tend to Gaussian processes (GPs) as their widths tend to infinity --- now you can play with these GP kernels in ! Try out RNN-GP, GRU-GP, Transformer-GP, or Batchnorm-GP today! Repo: Colab Entry Point:

    Poništi
  16. 27. pro 2019.

    It was a pleasure to visit the Physics department of Taiwan National University! What a beautiful place!

    Poništi
  17. 19. pro 2019.

    15/ The tensor program framework is very powerful and has many other consequences, such as the universality of Neural Tangent Kernels. If you would like to know more, check out

    Prikaži ovu nit
    Poništi
  18. 19. pro 2019.

    14/ Finally, we verify our theory is more and more accurate as the width of the network increases. Here we measure the relative frobenius norm of the empirical kernel to the infinite-width theoretical kernel, and we see that the deviation tends to 0 like 1/sqrt(width)

    Prikaži ovu nit
    Poništi
  19. 19. pro 2019.

    13/ These are examples of kernels corresponding to the infinite-width limit of other architectures: GRU, transformer, batchnorm.

    Prikaži ovu nit
    Poništi
  20. 19. pro 2019.

    12/ Again, see for an outline of the proof.

    Prikaži ovu nit
    Poništi
  21. 19. pro 2019.

    11B/ While at face value, this may seem limited, one can in fact express almost all modern architectures in this framework: resnet, transformer, LSTM, etc. Here are some simple examples of tensor programs expressing NN computation.

    Prikaži ovu nit
    Poništi

Čini se da učitavanje traje već neko vrijeme.

Twitter je možda preopterećen ili ima kratkotrajnih poteškoća u radu. Pokušajte ponovno ili potražite dodatne informacije u odjeljku Status Twittera.

    Možda bi vam se svidjelo i ovo:

    ·