Quantcast
Channel: Machine Learning
Viewing all 57517 articles
Browse latest View live

[D] Cremer et al. Inference Suboptimality in Variational Autoencoders - how was figure 2 computed?

$
0
0

Figure 2 in this paper shows the "True posterior" of a VAE.

How could they have computed this? Some of you experts probably have a strong guess.

I thought the motivation for VAE is because the posterior (among other things) is truly difficult to compute?

submitted by /u/AloneStretch
[link] [comments]

[D] Poisoning attacks against neural networks

$
0
0

I'm wondering what current approaches are towards this problem in an adversarial distributed ML context.

One paper I came across attempts to group gradients into "indicative features" that exhibit a certain distribution in a non-attack setting, and compare incoming gradients against this distribution and reject them if they appear to be significant outliers.

This approach seems like it would be difficult to make work in a scenario where a given "learner" (assuming some kind of distributed or peer-to-peer network where nodes learn collaboratively) only has access to a subset of a global dataset, and thus can't know what the expected distribution is for a gradient that was produced by training on data it has never seen.

I suppose this problem could maybe be attended to by polling peers to see if anyone else has knowledge about this kind of data?

Another approach I considered was simply rejecting proposed weight updates if the update would cause loss to go up. But again this seems like it would only really work if each node/learner has access to the same validation set that is representative of the global dataset. Would also be problematic in the sense that sometimes a slight loss increase is just a natural part of stochastic gradient descent.

In distributed ML settings, what are common approaches to defending against malicious nodes seeking to subvert the overall learning process via poisoning?

submitted by /u/ConfuciusBateman
[link] [comments]

How do you deal with floating point roundoff error?

$
0
0

I'm doing on classifying problem using CNN with 20+ layers.

Because the data type is double between 0 and 1 and the floating-point arithmetic is different across compilers/platforms, the final classification result often changes on different environments with same test data+weights.

Example)

Computer 1:

First 4 outputs of layer 0: 4.2956220815 6.1773069574 6.2690120863 6.2555422433

First 4 outputs of layer 1: 6.1773069574 6.2690120863 6.2884041050 6.3546746538

First 4 outputs of layer 2: -0.0675196870 -0.0275491092 -0.0170720917 0.6978492065

First 4 outputs of layer 3: 3.2372829955 2.9805280654 2.2781343339 2.5922840719

First 4 outputs of layer 4: -0.1213837868 -0.1071539059 -0.0187412153 -0.0778026944

Computer 2:

First 4 outputs of layer 0: 4.2956220815 6.1773069574 6.2690120863 6.2555422433

First 4 outputs of layer 1: 6.1773069574 6.2690120863 6.2884041050 6.3546746538

First 4 outputs of layer 2: -0.0675196870 -0.0275491092 -0.0170720917 0.6978492065

First 4 outputs of layer 3: 3.3214833420 2.9399862621 3.1782110457 3.6103477371

First 4 outputs of layer 4: -0.1531310157 -0.1546919095 -0.0831776014 0.1733128951

After passing 4~5th layer, the classification result become totally different, making the model frail and unreliable.

How could I deal with this kind of problems..?

submitted by /u/frozenca
[link] [comments]

[R] Gotta Learn Fast: A New Benchmark for Generalization in RL (OpenAI, PDF)

[R] The Tsetlin Machine - a new approach to ML - outperforms Neural Networks, SVMs, Random Forests, the Naive Bayes Classifier and Logistic Regression

[N] Human Hippocampal Neurogenesis Persists throughout Aging

[R]https://aeon.co/videos/a-neural-network-that-keeps-seeing-art-where-we-see-mundane-objects

[P] A birds-eye view of optimization algorithms


[P] Cuttlefish: A Lightweight Primitive for Adaptive Query Processing and Machine Learning

[News] Comet.ml - Automatically tracking of Machine Learning Experiments

$
0
0

Hi, We built comet.ml to allow machine learning engineers to automatically track their machine learning code, experiments, hyperparameters and results. We're officially out of beta and would love your feedback! Comet is free for public/open source projects and students. It's really easy to use and works no matter where you train your models.

I'm one of the founders so AMA!

http://www.comet.ml

submitted by /u/gidime
[link] [comments]

[D] Machine Learning Glossary | Google Developers

[D] Mobilenet v2 paper said Depthwise Separable convolution speedup conv op 8-9 times without reducing much accuracy. Should we just use it all the time now? Is there any detail analysis on it?

[P] Convolutional Generative Adversarial Network to Synthesize Novel Quilt Designs

[project] Check out this visualization of the Newton Raphson method created using my open source library - https://github.com/ryu577/pyray

[R] The Kanerva Machine: A Generative Distributed Memory


[P] Live stream of neural network learning to play Minecraft! (Behavioral cloning)

[R] [1204.5721] Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems

[D] Does anyone know any really good papers on spiking neural networks?

[D] Self-teaching more advanced math required for ML/DL research

$
0
0

I just started doing my Masters in AI this year after finishing a CS bachelor, and am intending to most likely continue on with a PhD in ML/DL (if I can). The problem is, now that I'm finally getting deep into studying ML/DL materials I'm finding that my math skills are really not good enough.

To give a specific example, I just started reading Machine Learning: A Probabilistic Perspective. Prior to picking this up, I also looked at the wildly praised Bishop's Pattern Recognition and the Elements of Statistical Learning, but it felt that the book I picked best matched the things I don't know about / want to learn.

But just as I'm working through the probability refresher chapter, I'm starting to feel my usual "I should know more about this" feeling. To be extra specific, just the mention of student's/laplace/gamma/beta distributions are something I never experienced in my previous studies. Or things like multi-variate distributions. Or even something like hypothesis testing is something that was almost untouched during my undergrad.

To give another example, I've been trying to catch up on the recent DL papers and most of the time I don't have trouble with the math, but last week I was reading the Wasserstein GAN paper and just a mention of Lipschitz continuity makes me feel like a first year undergrad.

There are many more things I could list, but you probably get the general idea.

To put this in more perspective, my undergrad was quite math-ish, yet there are things that clearly weren't covered, or they were and I have no intuition of them. There also aren't really any more courses I can take that teach more of the advanced stuff without actually taking classes with the math majors at a different faculty.

The reason I'm saying all of this, is that I can't really imagine myself finishing my thesis next year and going into a PhD where it might be expected of me to write a paper like the WGAN one.

I'm not sure how this compares to US, though I asked my professor at a deep learning seminar and he basically said that he knows of one university in Switzerland that teaches DL with lots of applied theory, and that otherwise there isn't really any course I can take to learn the more advanced stuff in probability / linear algebra / calculus.

For example, in my two probability courses we covered stuff like basic probability theory, and then moved on to markov chains, bits of queue theory, bits of poisson processes. But looking at the ML: A Prob Perspective book, it feels as if I had only taken the most basic probability theory. In Linear algebra, we did all the basic stuff, all the fun decompositions (SVD, QR, LU), eigenvalues, orthogonal matrices, etc., but we didn't touch any vector calculus, and I don't really have any intuition for any of the more advanced stuff, other than understanding the definitions/proofs. As for calculus, we kinda ended with Lagrange multipliers and did a bit about theory of metric spaces, but again, reading most of the papers, I feel like knowing what a Jacobian is is the only useful thing I know.

Any tips on how I should continue? What books to read? What online courses to take? What videos to watch?

submitted by /u/progfu
[link] [comments]

[D] Retro Contest | OpenAI

Viewing all 57517 articles
Browse latest View live




Latest Images