Timezone: »

Efficiently Learning One Hidden Layer ReLU Networks From Queries
Sitan Chen · Adam Klivans · Raghu Meka

Thu Dec 09 04:30 PM -- 06:00 PM (PST) @
While the problem of PAC learning neural networks from samples has received considerable attention in recent years, in certain settings like model extraction attacks, it is reasonable to imagine having more than just the ability to observe random labeled examples. Motivated by this, we consider the following problem: given \emph{black-box query access} to a neural network $F$, recover $F$ up to some error. Formally, we show that if $F$ is an arbitrary one hidden layer neural network with ReLU activations, there is an algorithm with query complexity and runtime polynomial in all parameters which outputs a network $F’$ achieving low square loss relative to $F$ with respect to the Gaussian measure. While a number of works in the security literature have proposed and empirically demonstrated the effectiveness of certain algorithms for this problem, ours is to the best of our knowledge the first provable guarantee in this vein.

Author Information

Sitan Chen (UC Berkeley)
Adam Klivans (UT Austin)
Raghu Meka (UCLA)

More from the Same Authors