Deep Models Under the GAN: Information Leakage from Collaborative Deep Learning
Deep Learning has recently become hugely popular in machine learning, providing significant improvements in classification accuracy in the presence of highly-structured and large databases. Researchers have also considered privacy implications of d…
Authors: Bril, Hitaj, Giuseppe Ateniese
Deep Models Under the GAN: Information Leakage from Collaborative Deep Learning Briland Hitaj ∗ Stevens Institute of T echnology bhitaj@stevens.edu Giuseppe Ateniese Stevens Institute of T echnology gatenies@stevens.edu Fernando Perez-Cruz Stevens Institute of T echnology fperezcr@stevens.edu ABSTRA CT Deep Learning has recently become hugely popular in machine learning for its ability to solve end-to-end learning systems, in which the features and the classiers are learned simultaneously , providing signicant impr ovements in classication accuracy in the presence of highly-structured and large databases. Its success is due to a combination of recent algorithmic break- throughs, increasingly pow erful computers, and access to signi- cant amounts of data. Researchers have also consider ed privacy implications of deep learning. Models are typically trained in a centralized manner with all the data being processed by the same training algorithm. If the data is a collection of users’ private data, including habits, personal pictures, geographical positions, interests, and more, the central- ized server will have access to sensitive information that could potentially be mishandled. T o tackle this problem, collaborative deep learning models have recently been proposed where parties locally train their deep learning structur es and only share a subset of the parameters in the attempt to keep their respective training sets private. Parameters can also be obfuscated via dierential pri- vacy (DP) to make information extraction even mor e challenging, as proposed by Shokri and Shmatikov at CCS’15. Unfortunately , we show that any privacy-preserving collabora- tive deep learning is susceptible to a pow erful attack that w e devise in this pap er . In particular , we show that a distributed, federated, or decentralized deep learning approach is fundamentally broken and does not protect the training sets of honest participants. The attack we dev eloped exploits the real-time nature of the learning process that allows the adversary to train a Generative Adversarial Network (GAN) that generates prototypical samples of the targeted training set that was meant to be private (the samples generated by the GAN are intended to come from the same distribution as the training data). Interestingly , we show that record-lev el dierential privacy applied to the shared parameters of the mo del, as suggested in pre vious work, is ineective (i.e., r ecord-level DP is not designed to address our attack). KEY W ORDS Collaborative learning; Security; Privacy; Deep learning It’s not who has the best algorithm that wins. It’s who has the most data. A ndrew Ng Self-taught Learning ∗ The author is also a PhD student at University of Rome - La Sapienza 1 IN TRODUCTION Deep Learning is a new branch of machine learning that makes use of neural networks, a concept which dates back to 1943 [ 49 ], to nd solutions for a variety of complex tasks. Neural netw orks were inspired by the way the human brain learns to show that distributed articial neural networks could also learn nontrivial tasks, even though curr ent architectures and learning procedures are far from brain-like behavior . Algorithmic breakthroughs, the feasibility of colle cting large amounts of data, and increasing computational p ower have con- tributed to the current popularity of neural networks, in particular with multiple (deep) hidden layers, that indeed have starte d to outperform previous state-of-the-art machine learning techniques [ 6 , 29 , 75 ]. Unlike conventional machine learning approaches, deep learning needs no feature engineering of inputs [ 45 ] since the model itself extracts rele vant features on its own and denes which fea- tures are rele vant for each problem [29, 45]. Deep learning models perform extremely w ell with correlated data, which contributed to substantial improvements in computer vision [ 47 ], image processing, video processing, face recognition [ 82 ], speech recognition [ 34 ], text-to-speech systems [ 64 ] and nat- ural language processing [ 2 , 15 , 90 ]. Deep learning has also been used as a component in more complex systems that are able to play games [ 33 , 42 , 57 , 60 ] or diagnose and classify diseases [ 16 , 18 , 26 ]. Howev er , there are se vere privacy implications associated with deep learning, as the traine d model incorporates essential informa- tion about the training set. It is relatively straightforward to extract sensitive information from a model [4, 27, 28]. Consider the following cases depicted in Figure 1, in which N users store local datasets of private information on their respective devices and would like to cooperate to build a common discrimina- tive machine. W e could build a classier by uploading all datasets into a single location (e.g., the cloud), as depicted in Figure 1 (a). A service operator trains the model on the combined datasets. This centralized approach is very eective since the model has access to all the data, but it’s not privacy-preserving since the operator has direct access to sensitive information. W e could also adopt a col- laborative learning algorithm, as illustrated in Figure 1 (b), where each participant trains a local mo del on his device and shares with the other users only a fraction of the parameters of the model. By collecting and exchanging these parameters, the service operator can create a trained model that is almost as accurate as a model built with a centralized approach. The de centralized approach is considered more privacy-friendly since datasets are not exposed directly . Also, it is shown experimentally to converge ev en in the case when only a small percentage of model parameters is shared (a) Centralize d Learning (b) Collaborative Learning Figure 1: T wo approaches for distributed deep learning. In (a), the red links show sharing of the data between the users and the ser ver . Only the ser ver can compromise the privacy of the data. In ( b), the red links show sharing of the mo del parameters. In this case a malicious user employing a GAN can de ceive any victim into releasing their private information. and/or when parameters ar e truncated and/or obfuscated via dier- ential privacy [ 77 ]. But it needs sev eral training passes through the data with users updating the parameters at each epoch. The Deep Learning community has recently proposed Gener- ative Adversarial Networks (GANs) [ 30 , 70 , 72 ], which are still being intensively developed [ 3 , 9 , 32 , 43 , 56 ]. The goal of GANs is not to classify images into dierent categories, but to generate similar-looking samples to those in the training set (ideally with the same distribution). More imp ortantly , GANs generate these samples without having access to the original samples. The GAN interacts only with the discriminative deep neural netw ork to learn the distribution of the data. In this paper , we devise a pow erful attack against collaborative deep learning using GANs. The result of the attack is that any user acting as an insider can infer sensitive information from a victim’s device. The attacker simply runs the collab orative learning algo- rithm and reconstructs sensitive information stored on the victim’s device. The attacker is also able to inuence the learning pr ocess and deceive the victim into releasing more detailed information. The attack works without compromising the service operator and even when model parameters are obfuscated via dierential privacy . As depicted in Figure 1(a), the centralized server is the only player that compromises the privacy of the data. While in Figure 1(b), we show that any user can intentionally compr omise any other user , making the distributed setting even more undesirable. Our main contribution is to propose and implement a novel class of active inference attacks on deep neural networks in a collabora- tive setting. Our method is more eective than existing black-box or white-box information extraction mechanisms. Namely , our contributions are: (1) W e devise a new attack on distributed deep learning based on GANs. GANs are typically used for implicit density esti- mation, and this, as far as w e know , is the rst application in which GANs are used maliciously . (2) Our attack is more generic and eective than current infor- mation extraction mechanisms. In particular , our approach can be employed against convolutional neural networks (CNN) which are notoriously dicult for model inversion attacks [78]. (3) W e introduce the notion of de ception in collaborative learn- ing, where the adversary deceives a victim into releasing more accurate information on sensitive data. (4) The attack we devise is also eective when parameters are obfuscated via dierential privacy . W e emphasize that it is not an attack against dierential privacy but only on its proposed use in collaborative deep learning. In practice, we show that dierentially private training as applie d in [ 77 ] and [ 1 ] (example/r ecord-level dierential privacy) is ineec- tive in a collaborative learning setting under our notion of privacy . 2 REMARKS W e devise a new attack that is more generic and eective than current information extraction mechanisms. It is based on Gen- erative Adversarial Networks (GANs), which were proposed for implicit density estimation [ 30 ]. The GAN, as detailed in Section 5, generates samples that appear to come from the training set, by pitting a generative deep neural network against a discriminative deep neural network. The generative learning is successful when- ever the discriminative model cannot determine whether samples come from the GAN or the training set. It is imp ortant to realize that b oth the discriminative and generative networks inuence each other , b ecause the discriminative algorithm tries to separate GAN-generated samples from real samples while the GAN tries to generate more realistic looking samples (ideally coming from the same distribution of the original data). The GAN never se es the actual training set, it only r elies on the information stored in the discriminative model. The pr ocess is similar to the facial composite imaging used by police to identify suspects, where a comp osite artist generates a sketch from an eyewitness discriminative descrip- tion of the face of the suspe ct. While the composite artist (GAN) has never seen the actual face, the nal image is based on the feedback from the eye witness. W e use GANs in a new way , since they are used to extract in- formation from honest victims in a collab orative deep learning framework. The GAN creates instances of a class that is supposed to be private. Our GAN-base d method works only during the train- ing phase in collaborative deep learning. Our attack is eective even against Convolutional Neural Networks which are notori- ously dicult to invert [ 78 ], or when parameters are obfuscated via dierential privacy with granularity set at the record level ( as proposed in [ 77 ] and [ 1 ]). It works in a white-box access model where the attacker sees and uses internal parameters of the model. This in contrast to black-box access where the attacker sees only the output of the model for each particular input. It is not a limitation of our procedure because the purpose of collaborative learning is to share parameters, even if in a small percentage . Once the distribute d learning process ends, a participant can always apply a mo del inversion or similar attack to the trained model. This is not surprising. What we show in this paper is that a malicious participant can see how the mo del evolves and inuence other honest participants and force them into releasing relevant information about their private datasets. This ability to deceive honest users is unique to our attack. Furthermore, truncating or obfuscating shared parameters will not help since our attack is eective as long as the accuracy of the local models is high enough. W e emphasize however that our attack does not violate dier- ential privacy (DP) , which was dened to protect databases. The issue is that, in collaborative deep learning, DP is b eing applied to the parameters of the model and with granularity set at the record/example lev el. However , the noise added to learning param- eters will ultimately have to be containe d once the model b ecomes accurate. Our attack works whenever the model can accurately classify a class and will generate representatives of that class. The way DP is applied in [ 77 ] and [ 1 ] can at best protect against the re- covery of specic elements asso ciated with a label that was indeed used during the learning phase. The results of our attack may or may not be regarded as privacy violations. Consider the following examples: (1) The victim’s device contains standard medical records. The GAN will generate elements that look like generic medical records, i.e., items fr om the same distribution of those in the training set. The attacker may learn nothing of interest in this case, and there is no privacy violation. Ho wever , if the victim’s device contains records of patients with cancer then the attacker may see inexistent patients, but all with cancer . Depending on the context, this may be considered a privacy violation. (2) The victim’s device contains pornographic images. The GAN will generate similar scenes. While they may appear simu- lated, the information leaked to the adversar y is signicant. In other cases, our attack could b e useful to law enforcement ocials acting as adversaries. For instance, when the vic- tim’s device contains pedo-pornographic images or training material for terrorists. (3) The victim’s device contains speech recordings. The GAN will generate babbling, with lots of ctitious word-like sounds (comparable to W aveNet [ 64 ] when the network is trained without the text sequence), thus there is no privacy viola- tion. However , it may be p ossible to infer the language used (e.g., English or Chinese) or whether the speaker is male or female, and this leake d information may constitute a privacy violation. (4) The victim’s device contains images of Alice. The GAN will generate faces that resemble Alice much like a composite artist generates a sketch of an eyewitness’s memory of Alice. In our attack framework, the adversar y will also collect all these drawings of Alice and falsely claim they are Ev e’s. This will force the local model within the victim’s device to re- lease more r elevant and distinctive details ab out Alice ’s face, exacerbating the leakage . However , while many see this as a privacy violation, others may disagree since the adversar y may not recover the exact face of Alice but only a reconstruc- tion (see Figure 2) . On the other hand, if Alice wears glasses or has brown hair , then this information will be leaked and may constitute a privacy violation depending on the context. A further example is given in Figure 3, wher e DCGAN was run on the CIF AR-10 dataset [ 41 ] while targeting a class consisting of approximately 6,000 images containing various horses. Note that the class could be labele d ‘jj3h221f ’ and make no obvious reference to horses. The images produced by the GAN will tell the adversary that class ‘jj3h221f ’ does not contain cars or airplanes but animals (likely horses). Dierential privacy in collaborative learning is meant to protect the recovery of specic elements used during training. Namely , an adversary cannot tell whether a certain X was included in the training set (up to a certain threshold value). W e circumvent this protection by generating an X ′ which is indistinguishable from X . In Figure 2, we show a real example of a face X along with X ′ , the image generated by the GAN. Both images look similar even though X ′ is not X . While this does not violate DP, it clearly leads to severe privacy violations in many cases. Our point is that example/record- level DP is inadequate in this context, much like se cure encryption against a chosen-plaintext attack (CP A) is inadequate in an active adversarial environment. Ther e is nothing wrong with DP per se (as there is nothing wrong with CP A -secure encryption); clearly DP provides information-theoretic protection but it’s important to set its level of granularity right. At record level, it is just not enough to protect sensitive information in collaborative learning against active adversaries. One can consider DP at dierent granularities (e .g., at user or device le vel) but this is not what is proposed in [ 77 ]. Researchers can keep arguing about the proper use of DP or what DP is supp osed to protect [ 40 , 53 , 54 , 58 ], but ultimately , in the context of this work, one should ask: W ould I use a system that let casual users recover images that are eectively indistinguishable from the ones in my picture folder? The p oint is that collab orative learning for privacy is less desirable than the centralized learning approach it was supposed to improve upon: In centralize d learning only the ser vice provider can violate users’ privacy , but in collab orative learning, any user may violate the privacy of other users in the system, without involving the service provider (see Figure 1). Figure 2: Picture of Alice on the victim’s phone, X , and its GAN reconstruction, X ′ . Note that X ′ , X , and X ′ was not in the training set. But X ′ is essentially indistinguishable from X . Figure 3: GAN-generated samples for the ‘horse’ class from the CIF AR-10 dataset 3 IMP ACT Google adopts a centralized approach and collects usage informa- tion from Android devices into a centralized database and runs machine learning algorithms on it. Google has recently introduced Federated Learning [ 50 , 51 ] to enable mobile de vices to collabora- tively learn a shared prediction model while keeping all the train- ing data local. Devices download the current model from a Google server and improve it by learning from local data. Federated learning appears to b e the same as collaborative learn- ing, and our attack should b e e qually eective. In the end, each device will download the trained mo del from the Go ogle server , and the GAN will be able to operate successfully as long as the lo cal model is learning. In federated learning, it is possible to protect individual model updates. Rather than using dierential privacy as in [ 77 ], Google proposes to use a secure aggr egation protocol. The updates from individual users’ devices are se curely aggregated by leveraging secure multiparty computation (MPC) to compute weighted av er- ages of model parameters [ 8 ] so that the Google server can decrypt the result only if several users have participated. W e believe that this mechanism, as described in their paper , is ineective against our attack architecture since we simply r ely on the fact that local models have successfully learned. Their security model considers only the case in which Google is the adversary that scrutinizes individual updates. Therefore, they don’t consider the point we raise in this paper that casual users can attack other users. This makes fe derated learning p otentially even more dangerous than the centralized one it is supposed to replace, at least in its current form. Indeed, our assessment is based on the description given in an announcement and two resear ch papers. W e have had no access to the actual implementation of the system yet, and products tend to improve signicantly o ver time. Apple is said to apply dierential privacy within a crowdsourced learning framework in future versions of iOS [ 35 ]. While we do not know the details, we hope our pap er ser ves as a warning on the risks of applying dierential privacy improperly in collaborative deep learning. Our adversary does not have to work for the service provider , but he is a regular user targeting another user , e.g., a celebrity or a politician. 4 RELA TED W ORK Deep Learning has proven to be successful in various areas of com- puter science. The capability to learn, process and produce relevant information from large quantities of data, makes deep learning a good option for the cyber security domain as well. Howev er , new and unique attacks have emerged that pose a serious threat to the privacy of the information being processed. 4.1 Attacks on Machine Learning Mo dels T o the best of our kno wledge, the rst w ork that deals with extract- ing unexpected information from trained models is the one from Ateniese et al. [ 4 ] (released in 2011 and on arXiv in 2013 [ 4 , 12 ]). There, the authors devised a meta-classier that is trained to hack into other machine learning classiers to infer sensitive information or patterns fr om the training set. For instance, they were able to ex- tract ethnicity or gender information from trained voice recognition systems. The work was later extended by Fredrikson et al. [ 27 , 28 ] where they proposed model inversion attacks on machine learning algo- rithms by exploiting condence information re vealed by the model. For instance , when applied to facial recognition systems, they show that it is possible to reconstruct images ab out a particular label known to the adversary . Recently , the work of Tramèr et al. [ 83 ] shows that stealing machine learning models is p ossible when taking into consideration only the predictions provided by the mo del. Membership inference attacks were developed by Shokri et al. [ 78 ]. Here, the adversary is given black-box access to the mo del and can infer whether a certain record was originally in the training set. McPherson et al. [ 52 ] use deep learning to infer and re veal the identity of subjects b ehind blurred images. In their work, Papernot et al. [ 66 ] show that an adversarially crafted input can be fed to deep learning models and make them prone to error , i.e., make the model misclassify the input therefore producing incorrect outputs. For example, a STOP sign on the road can b e subtly mo died to look the same to human eyes, but that is classied as another sign by a trained model. The work was extended in [36, 44, 65, 87]. 4.2 Privacy Preser ving Machine Learning Defense me chanisms against p owerful adversaries were devised by Shokri and Shmatikov [ 77 ]. The authors introduce the concept of distributed deep learning as a way to protect the privacy of training data [ 85 ]. In this model, multiple entities collaboratively train a model by sharing gradients of their individual mo dels with each other through a parameter ser ver . Distribute d learning is also considered in [ 17 , 51 , 59 , 80 , 89 , 91 ]. Mohassel et al. [ 61 ] provide a so- lution for training neural networks while preser ving the privacy of the participants. Howev er , it deploys secure multiparty computation in the two-server model where clients outsource the computation to two untrusted but non-colluding servers. Howev er , Shokri and Shmatikov [ 77 ] are the rst to consider privacy-preser ving measures with the purpose of nding practical alternatives to costly multi-party computation (MPC) techniques . Google developed techniques to train models on smartphones directly without transferring sensitive data to the company’s data centers [ 8 , 51 ]. Microsoft developed Cr yptoNets [ 20 ] to perform deep learning on encrypte d data and provide encrypte d outputs to the users [ 86 ]. Ohrimenko et al. [ 63 ] developed data-oblivious machine learning algorithms trained on trusted processors. Dier- ential privacy plays an imp ortant role in de ep learning as shown in [1, 39, 77, 79]. 4.3 Dierential Privacy Dierential Privacy (DP) was introduce d by Dwork [ 21 ]. Its aim is to provide pro vable privacy guarantees for database records without signicant query accuracy loss. Dierential privacy for big data was considered by Dwork et al. [ 23 ]. Several w orks have adopted DP as an ecient defense me chanism [ 5 , 7 , 11 , 13 , 19 , 24 , 25 , 38 , 55 , 62, 67, 74, 88]. Collaborative deep learning proposed by Shokri and Shmatikov [ 77 ] uses DP to obfuscate shared parameters while Abadi et al. [ 1 ] propose to apply DP to the parameters during training. DP was used in deep auto-encoders in [69]. Covert channels, ho wever , can be used to defeat DP-protected databases as shown in the work of Haeberlen et al. [ 37 ]. In general, privacy cannot be guaranteed if auxiliary information (outside the DP model) is accessible to the adversar y [ 22 ]. At NDSS’16, it was shown by Liu et al. [ 48 ] that DP at a certain granularity is not eective in real-life scenarios where data such as so cial data, mobile data, or medical records have strong corr elations with each other . Note that it’s a matter of setting DP granularity right and DP is not being violated at all. 4.4 Privacy-Preserving Collaborative Deep Learning A centralized approach to deep learning forces multiple participants to pool their datasets into a large central training set on which it is possible to train a model. This poses serious privacy threats, as pointed out by Shokri and Shmatikov [ 77 ], and distrustful partici- pants may not be willing to collaborate. Considering the security and privacy issues described above, Shokri and Shmatikov [ 77 ] introduce a new collaborative learning approach, which allows participants to train their models, without explicitly sharing their training data. They exploit the fact that op- timization algorithms, such as Stochastic Gradient Descent (SGD), can be parallelized and executed asynchronously . Their approach includes a selective parameter sharing pr ocess combined with lo- cal parameter updates during SGD . The participants share only a fraction of their local mo del gradients through a Parameter Server (PS). Each participant takes turns and uploads and downloads a percentage of the most recent gradients to avoid getting stuck into local minima. This process only works if the participants agree in advance on a network architecture [77]. It is possible to blur the parameters shared with PS in various ways. Other than just uploading a small percentage of all the gradi- ents, a participant can also select certain parameters that are above a threshold, within a certain range, or noisy in agreement with dierential privacy procedures. 5 BA CK GROUND Supervise d machine learning algorithms take lab eled data and pro- duce a classier ( or regressor) that it is able to accurately predict the label of new instances that has not seen before. Machine learning algorithms follow the inductive learning principle [ 84 ], in which they go from a set examples to a general rule that works for any data coming from the same distribution as the training set. Given independent and identically distributed (i.i.d.) samples from p ( x , y ) , i.e., D = { x i , y i } n i = 1 , where x i ∈ R d and y i ∈ { 1 , 2 , . . . } , they solve the following optimization problem to nd an accurate classier: b θ = arg min θ ∈ Θ Õ i L ( f ( x i ; θ ) , y i ) + Ω ( θ ) , (1) where ˆ y = f ( x ; b θ ) represents the learning machine, i.e., for any input x it provides an estimate for the class label y . L ( w , y ) is a loss function that measures the error for misclassifying y by w . And Ω ( θ ) is a regularizer (independent of the training data) that avoids overtting. Sup ervised learning algorithms like Support V e ctor Machines (SVMs) [ 76 ], Random Forests [ 10 ], Gaussian Processes (GPs) [ 71 ] and, of course , deep neural networks [ 29 ] can b e depicted by this general framework. Deep neural networks ar e becoming the weapon of choice when solving machine-learning problems for large databases with high- dimensional strongly correlated inputs be cause they are able to provide signicant accuracy gains. Their impr ovements are based on additionally learning the features that go into the classier . Be- fore deep learning, in problems that dealt with high-dimensional strongly correlated inputs (e.g., images or voice), humanly engi- neered features, which were built to reduce dimensionality and correlation, were fed to a classier of choice . The deep neural net- work revolution has shown that the features should not be humanly engineered but learned from the data , because the hand-code d fea- tures were missing out relevant information to produce optimal results for the available data. The deep neural network learns the useful featur es that make sense for each pr oblem, instead of relying on best guesses. The deep neural network structures are designed to exploit the correlation in the input to learn the features that are ideal for optimal classication. The deep structure is nee ded to extract those features in several stages, moving from local features in the lower layers to global features at the higher layers, b efore providing an accurate prediction on the top layer . These results have be come self-evident when datasets have grown in size and richness. The learning machine f ( x ; θ ) summarizes the training database in the estimated parameters b θ . From the learning machine and its estimated parameters, relevant features of the training database, if not complete training examples, can be recov ered. So an adver- sary that wants to learn features from the original training data can do so if it has access to the learning machine. For example, SVMs store prototypical examples from each class in b θ and GPs store all the training points, so there is no challenge there for an adversary to learn prototypical examples for each class in those classiers. For deep neural networks, the r elation between b θ and the training points in D is more subtle, so researchers hav e tried to show that privacy is a possibility in these networks [ 77 ]. But the model inversion attack [ 27 , 28 ] has proven that we can recover inputs (e.g., images) that look similar to those in the training set, leaking information to the adversary about how each class lo oks like. And as deep neural networks are trained with unprocessed inputs, these attacks recover prototypical e xamples of the original inputs. It is important to emphasize that this is an intrinsic property of any machine-learning algorithm. If the algorithm has learned and it is providing accurate classication, then an adversary with access to the model can obtain information from the classes. If the adversary has access to the model, it can recover prototypical examples from each class. If sensitive or private information is needed for the classier to perform optimally , the learning machine can potentially leak that information to the adversary . W e cannot have it both ways, either the learning machine learns successfully , or data is kept private. 5.1 Limitations of the Mo del Inversion Attack The model inversion attack works in a simple way [ 27 , 28 ]: Once the network has been trained, we can follow the gradient used to adjust the weights of the network and obtain a reverse-engineered example for all represented classes in the netw ork. For those classes that we did not have prior information, we would still be able to recover prototypical examples. This attack shows that any accurate deep learning machine, no matter how it has been trained, can leak information about the dierent classes that it can distinguish. Moreover , the model inversion attack may recover only proto- typical examples that have little resemblance to the actual data that dened that class. This is due to the rich structure of deep learning machines, in which broad areas of the input space are classied with high accuracy but something else is left out [ 31 , 81 ]. If this is the case, the adversary might think he has recovered sensitive information for that class when he is just getting meaningless in- formation. For example, w e refer the reader to Figure 5 from [ 81 ], where six training images for a scho ol bus, bird, a temple, soap dispenser , a mantis and a dog have be en slightly tweaked to be classied as an ostrich (Struthio camelus), while they still look like the original image. In [ 31 ], the authors sho w in Figure 5 a procedure similar to the model inversion attack. A randomly generated image , plus gradient information from the de ep belief netw ork, produces a random looking image that is classied as an airplane. The structure of deep neural networks is so large and e xible that it can be fooled into giving an accurate label even though the image to a human looks nothing like it. Thus any model inversion attack can obtain private information from a trained de ep neural network, but it can land in an unrep- resented part of the input space that looks nothing like the true inputs dened for each class. Extensive resear ch in the ML commu- nity has shown that GAN generated samples are quite similar to the training data, thus the results coming from our attack reveal more sensitive information ab out the training data compared to the average samples or aggregated information one w ould expect from a model inversion type of attack. 5.2 Generative Adversarial Networks One way to address the problem highlighted in [ 31 , 81 ] is generating more training images so to cover a larger p ortion of the space. This can be accomplished through Generative Adversarial Networks (GANs) [30]. The GAN pr oce dure pits a discriminative deep learning network against a generative deep learning network. In the original paper [30], the discriminative network is trained to distinguish between images from an original database and those generated by the GAN. The generative network is rst initialized with random noise , and at each iteration, it is trained to mimic the images in the training set of the discriminative netw ork. The optimization problem solv ed by the GAN procedure can be summarized as min θ G max θ D n + Õ i = 1 log f ( x i ; θ D ) + n − Õ j = 1 log ( 1 − f ( д ( z j ; θ G ) ; θ D )) (2) where x i are images from the original data and z j are randomly generated images (e.g., each pixel distributed between 0 and 255 uniformly). Let f ( x ; θ D ) be a discriminative deep neural network that, given an image, produces a class lab el and let θ D denote its parameters. Let д ( z ; θ G ) be a generative de ep neural network, which given a random input produces an image. The training procedure works as follows. First, we compute the gradient on θ D to maximize the performance of the discriminative deep neural network. Hence f ( x ; θ D ) is able to distinguish between samples from the original data, i.e., x i , and samples generated from the generativ e structure , i.e., x fake j = д ( z j ; θ G ) . Second, we compute the gradients on θ G , so the samples generate d from x fake j = д ( z j ; θ G ) look like a perfect replica of the original data 1 . The procedure ends when the discriminative network is unable to distinguish between samples from the original database and the samples generated by the generative network. The authors of the paper [30] prove the following theorem: Theorem 5.1. The global minimum of the virtual training criterion in (2) is achieved if and only if p ( x ) = p ( д ( z ; θ G )) . The theorem shows that the adversarial game ends when the GAN is generating images that appear to come fr om the original dataset. In [ 32 ], the author shows that in the innite sample limit the generative network would draw samples from the original training 1 The generated data looks like the original data, but they are not copies of them. distribution. But it also recognizes that the GAN procedure will not converge. In a recent paper [ 72 ], the authors have signicantly im- proved the training of the GAN including new features to improve convergence to the density model. 6 THREA T MODEL Our threat model follows [77], but relies on an active insider . The adversary pretends to be an honest participant in the col- laborative deep learning protocol but tries to extract information about a class of data he do es not own. The adversary will also sur- reptitiously inuence the learning process to deceive a victim into releasing further details about the targeted class. This adversarial inuence is what makes our attack mor e eective than, for instance, just applying model inversion attacks [ 27 ] against the nal trained model. Furthermore, our attack works for more general learning models (those for which a GAN can b e implemented), including those on which model inversion attack is notoriously inee ctive (e.g., conv olutional neural networks). Specically , we consider the following scenario: • The adversary works as an insider within the privacy-preserving collaborative deep learning protocol. • The objective of the adversary is to infer meaningful infor- mation about a label that he does not own. • The adversary does not compromise the central parameter server (PS) that colle cts and distributes parameters to the participants. That is, the parameter server , or the ser vice provider in our example, is not under the control of the adversary . In our real-world example, the adversary is a full- edged insider and does not have to work for the service provider . • The adversary is active since he directly manipulates values and builds a GAN locally . At the same time, he follows the protocol specication as viewed by his victims. In particular , the adversary takes turns, follows the parameter selection procedures, uploads and downloads the corr ect amount of gradients as agreed in advance, and obfuscates the uploaded parameters as required by the collaborative learning pr ocess. • As in [ 77 ], it is assumed that all participants agree in advance on a common learning objective . This implies that the adver- sary has kno wledge of the model structure and, in particular , of the data labels of other participants. • Unlike static adversaries as in model inversion [ 27 ], our ad- versary is allowed to be adaptive and work in real time while the learning is in progress. The adversary will b e able to inuence other participants by sharing specially-crafted gra- dients and trick participants into leaking more information on their local data. This is possible because the distributed learning procedure needs to run for several rounds before it is successful. 7 PROPOSED A T T A CK The adversary A participates in the collaborative deep learning protocol. All participants agree in advance on a common learning objective [ 77 ] which means that they agr ee on the type of neural network architecture and on the labels on which the training w ould take place. Let V be another participant (the victim) that declares labels [ a , b ] . The adversary A declares labels [ b , c ] . Thus, while b is in common, A has no information about the class a . The goal of the adversary is to infer as much useful information as possible about elements in a . Our insider employs a GAN to generate instances that look like the samples from class a of the victim. The insider inje cts these fake samples from a , as class c into the distributed learning procedure. In this way , the victim ne eds to work har der to distinguish between classes a and c and hence will reveal more information about class a than initially intended. Thus, the insider mimics samples from a and uses the victim to improve his knowledge ab out a class he ignored b efore training. GANs were initially devised for density estimation, so we could learn the distribution of the data from the output of a classier without seeing the data directly . In this case, we use this property to deceive the victim into providing more information about a class that is unknown to the insider . For simplicity , we consider rst two players (the adv ersar y and the victim) and then extend our attack strategy to account for multiple users. Each player can declare any number of labels, and there is no need for the classes to overlap. (1) Assume two participants A and V . Establish and agree on the common learning structure and goal. (2) V declares labels [ a , b ] and A lab els [ b , c ] . (3) Run the collab orative de ep learning protocol for several epochs and stop only when the model at the parameter server (PS) and both local mo dels have reached an accuracy that is higher than a certain threshold. (4) First, the Victim trains the network: (a) V downloads a percentage of parameters from PS and updates his local model. (b) V ’s local mo del is trained on [ a , b ] . (c) V uploads a selection of the parameters of his lo cal model to PS. (5) Second, the Adversary trains the network: (a) A downloads a percentage of parameters from the PS and update his local model. (b) A trains his local generative adversarial network (unknown to the victim) to mimic class a from the victim. (c) A generates samples from the GAN and lab els them as class c . (d) A ’s local model is traine d on [ b , c ] . (e) A uploads a selection of the parameters of his local model to PS. (6) Iterate between 4) and 5) until convergence. The steps highlighted in 5b) and 5c) above r epresent the extra work the adversary perform to learn as much as p ossible elements of the targeted label a . The procedure is depicted in Figure 4. The generalization of the attack to multiple users is reporte d in Algo- rithm 1. The GAN attack works as long as A ’s local model improves its accuracy over time. Another important point is that the GAN attack works even when dierential privacy or other obfuscation techniques are employed. It is not an attack on dierential privacy but on its propose d use in collab orative deep learning. Though there might be a degradation in the quality of results obtained, our Figure 4: GAN Attack on collaborative deep learning. The victim on the right trains the model with images of 3s (class a ) and images of 1s (class b ). The adversary only has images of class b (1s) and uses its label c and a GAN to fool the victim into releasing information about class a . The attack can b e easily generalized to several classes and users. The adversary does not even nee d to start with any true samples. experiments show that as long as the model is learning, the GAN can improve and learn, too. Of course, there may always exist a setup where the attack may be thwarted. This may b e achieved by setting stronger privacy guarantees, releasing fewer parameters, or establishing tighter thresholds. How ever , as also shown by the results in [ 77 ], such measures lead to models that are unable to learn or that perform worse than models trained on centralized data. In the end, the attack is eective even when dierential privacy is deployed, because the success of the generativ e-discriminative synergistic learning relies only on the accuracy of the discriminative model and not on its actual gradient values. 8 EXPERIMEN T AL SET UP The authors of [ 77 ] provided us with their source code that imple- ments a complete distributed collaborative learning system. Our attacks were run using their implementation of dierential privacy . 8.1 Datasets W e conducte d our experiments on two well-known datasets, namely MNIST [ 46 ] and A T&T dataset of faces [ 73 ] (a.k.a. Olivetti dataset of faces). 8.1.1 MNIST Dataset of Images. MNIST is the b enchmark dataset of choice in several deep learning applications. It consists of hand- written grayscale images of digits ranging from 0 to 9. Each image is of 32 × 32 pixels and centered. The dataset consists of 60,000 training data records and 10,000 records serving as test data. 8.1.2 A T&T Dataset of Faces (Olivei dataset). A T&T dataset, previously used also in the w ork of [ 27 ], consists of grayscale im- ages of faces of several persons taken in dierent positions. The version used in our experiments consists of 400 images of 64 × 64 pixels. 2 The dataset contains images of 40 dierent persons, namely 10 images per person. 2 http://www.cs.nyu.edu/ ˜ roweis/data.html For these experiments, we did not conduct any pre-processing of the data. The only pr ocessing performed on the data was scaling every image to the [− 1 , + 1 ] range, similar to [ 70 ]. This was done to adopt the state-of-the-art generator model of [ 70 ], which has a hyperbolic tangent t anh activation function in its last layer , thus outputting results in the [− 1 , + 1 ] range as well. 8.2 Framework W e build our experiments on the T orch7 scientic computing frame- work. 3 T orch is one of the most widely used deep learning frame- works. It provides fast and ecient construction of deep learning models thanks to LuaJI T 4 , a scripting language which is based on Lua 5 . 8.3 System Architecture W e use d a Convolutional Neural Network (CNN) based architecture during our experiments on MNIST and A T&T . The layers of the networks are sequentially attached to one another based on the nn . S equ ent i al () container so that layers are in a fee d-forward fully connected manner . 6 In the case of MNIST (Figure 15), the model consists of two convolution layers, nn . Sp at i al C on v ol u t i on M M () , where the t anh function is applied to the output of each layer before it is for warded to the max pooling layers, nn . Sp at i al M ax P ool i n д () . The rst con- volutional layer has a convolution kernel of size 5 × 5 and it takes one input plane and it produces 32 output planes. Whereas the second convolutional layer takes 32 input planes and produces 64 output planes and it has a convolution kernel of size 5 × 5. After the last max pooling layer , the data gets reshaped on a tensor of size 3 http://torch.ch/ 4 http://luajit.org 5 https://www.lua.org 6 https://github.com/torch/nn/blob/master/doc/containers.md#nn.Sequential Algorithm 1 Collaborative Training under GAN attack Pre- T raining Phase: Participants agree in advance on the follow- ing, as pointed out also by [77]: (1) common learning architecture, (model, labels etc.) {For ex. V declares labels [ a , b ] and A labels [ b , c ] } (2) learning rate, (lr) (3) parameter upload fraction (percentage), ( θ u ) (4) parameter download fraction, ( θ d ) (5) threshold for gradient selection, ( τ ) (6) bound of shared gradients, ( γ ) (7) training procedure, (sequential, asynchronous) (8) parameter upload criteria {cf. [77]} Training Phase 1: for epoc h = 1 to n r Epoc hs do 2: Enable user X for training 3: User x downloads θ d parameters from PS 4: Replace respective lo cal parameters on user x local model with newly downloaded ones 5: if ( us e r _ t y p e == AD V E RS ARY ) then 6: Create a replica of local f r e s hl y upd at e d model as D (dis- criminator) 7: Run Generator G on D targeting class a (unknown to the adversary) 8: Update G base d on the answer from D 9: Get n-samples of class a generated by G 10: Assign label c (fake label) to generate d samples of class a 11: Merge the generate d data with the local dataset of the adversary 12: end if 13: Run SGD on local dataset and update the local mo del 14: Compute the gradient vector ( new P ar amet e r s − ol d P ar amet e r s ) 15: Upload θ u parameters to PS 16: end for 17: return Collaboratively Trained Model {At the end of training, the adversary will have prototypical e xamples of members of class a known only to the victim} 256, on which a linear transformation is applied which takes as input the tensor of size 256 and outputs a tensor of size 200. Then a t anh activation function is applied to the output, which is then followed by another linear transformation which takes as input the tensor of size 200 and outputs a tensor of size 11. W e mo dify the output layer from 10 to 11, where the 11th output is where the adversary trains with the results generate d by G . As in Goodfellow et. al [ 30 ], the 11th class is the class where the ‘fake’ images are placed. Further details are provided on Section 9. The last layer of the models is a LogSoftMax layer , nn . Lo д So f t M ax () . Images in the A T&T dataset of faces are larger (64 × 64). Therefore, we built a convolutional neural network (Figure 17) consisting of three convolution layers and thr ee max pooling layers, followed by the fully connected lay ers in the end. As in the MNIST ar chitecture, t anh is used as an activation function. This mo del has an output layer of size 41, namely 40 for the real data of the persons and 1 as the class where the adversary puts the reconstructions for his class of interest. Since faces are harder to reconstruct than numbers, we implemented Algorithm 1 dierently . For this case, the generator G queries the discriminator D more times per epoch (size of adversary’s training data divided by batch size) to improve faster . The G ene r at or ( G ) architecture used in MNIST -related experi- ments, Figure 16, consisted of 4 convolution layers corresponding to nn . Sp at i al F ul l C on v olu t ion () from the torch ‘nn’ library . Batch normalization, nn.SpatialBatchNormalization() , is applied to the output of all layers e xcept the last one. The activation function is the rectied linear unit function, nn . Re L U () . The last layer of the model is a hyperbolic tangent function, t anh , to set the output of G to the [ -1, +1] range. Since A T&T images are larger (64x64), G has an additional (5th) convolution layer . The number of convolution layers needed were computed automatically using the techniques from [ 68 ]. G takes as input a 100-dimensional uniform distribution [ 14 , 70 ], and converts it to a 32x32 image for MNIST or a 64x64 image for A T&T . A s in [ 14 ], we initialized the weights of the gener- ator with 0 mean and 0.02 standard deviation. While [ 70 ] applies this initialization function to both D and G , w e do it only to G since D is the mo del that is shared among all participants. Both architectures described above are represented in Figure 16 and 18 as printed out by T orch7. W e refer the reader to Appendix A for further details on the architectures provided by T orch7. 8.4 Hyperparameter Setup For the MNIST -related experiments, we set the learning rate for both the collaboratively traine d model and the discriminator model to 1 e − 3 , learning rate decay of 1 e − 7 , momentum 0 and batch size of 64 . For the A T&T -related experiments, we set the learning rate to 0 . 02 and a batch size of 32 . Whereas, for the A T&T experiments concerning the multi-participant scenario, we used a batch size of 1 . W e kept the rest of the hyperparameters similar to the MNIST case. A learning rate of 0 . 02 worked better as it allowed more stochasticity in the process, thus allowing the model to converge faster . The authors of DCGAN [ 70 ] use the A d a m optimizer with a learn- ing rate of 0 . 0002 and a momentum term β 1 of 0.5 as pro vided in the tor ch implementation of DCGAN [ 14 ]. W e modie d the process to use sto chastic gradient descent (SGD) and, for this conguration, a learning rate of 0 . 02 for the generator worked better . 9 EXPERIMEN TS W e now evaluate how well our GAN procedure can recover r ecords from other participants. W e focus our experiments on MNIST and A T&T datasets that contain images. In principle, however , our ad- versarial strategy can be extended to other typ es of data, such as music, audio, medical records, etc. W e rst compar e our GAN at- tack against model inversion in a traditional setting. As mentioned before, model inversion has several limitations and may not be ef- fective against certain types of neural networks. While this may be clear from a theoretical perspective, we also provide experimental evidence for this claim in the rst experiment. Actual Image MIA DCGAN Figure 5: Results obtained when running mo del inversion attack (MIA) and a generative adversarial network (DCGAN) on CNN trained on the MNIST dataset. MIA fails to produce clear results, while DCGAN is successful. In the second set of experiments, we show how the GAN attack also works in the distributed setting in which the adversary is oblivious to the content of some, or all, labels, see Figure 7. In the third set of experiments, we show that adding noise to the parameters of the deep neural network before they are uploaded to the parameter server does not protect against our GAN attack. In general, deploying record-lev el dierential privacy to obfuscate the model parameters is inee ctive against our attack. The ecacy of the GAN is only limited by the accuracy of the discriminator . 9.1 MI Attack vs. GAN Attack In this rst example, we compare the model inversion (MI) and the GAN attacks, and we provide them with all the data. The adversar y has access to the fully trained models. For the MI attack, we train a convolutional neural network on all 60,000 training examples of the MNIST dataset. W e apply the model inversion attack in [ 27 ], once the deep neural network is trained. However , instead of approximating the derivatives as in [ 27 ], we collected the exact gradients computed by the model on the input given and the label (class) of interest. The results are shown in Figure 5. MI works well for MLP networks but clearly fails with CNNs. This is consistent with the work [ 78 ] where the authors attained similar results. It appears that MI is not eective when dealing with more complicate d learning structures. While relevant information is in the network, the gradients might take us to an area of the input space that is not repr esentative of the data that we are trying to recover . For the GAN approach, we adopt the DCGAN architecture in [ 70 ], and its torch implementation from [ 14 ]. The model consists of the discriminator (D) in combination with the DCGAN generator (G). W e made the generator model compatible with MNIST -type of images and use d metho ds propose d in [ 68 ] so that our code could automatically calculate the numb er of convolution layers needed. W e refer the reader to Section 8.3 for further details on the architectures. W e ran the experiments 10-times ( once per each class present in the MNIST dataset), and we let the models train until the accuracy reached by D was above 97% . W e show the results in Figure 5. Note a signicant dierence: In the GAN attack, the generative model is trained together with the discriminative model, while in MI, the discriminative mo del is only accesse d at the end of the training phase. Howe ver , this type of real-time access to the model is what makes our attack applicable to collab orative deep learning. 9.2 GAN Attack on Collab orative Learning without Dierential Privacy Now w e set the GAN attack in a collab orative envir onment like the one proposed in [ 77 ]. W e use the model describ ed in Section 7 and depicted in Figure 4. 9.2.1 Experiments on MNIST . Instead of using two labels per user , we use ve lab els for the rst user and six labels for the second user . The rst user has access to images of 0 to 4 (with label 1 to 5) and the second user , the adversary , has access to images of 5 to 9 ( label 6 to 10). The adversary uses its sixth class to extract information on one of the labels of the rst user . The results are shown in Figure 6. For every retrieved image (bottom row), we placed above it an actual training image from the rst user (we show the image that is closest in L1-norm). W e have repeated the experiment with three dierent parameter settings. In (a), the users upload and download the entire model. In ( b), the users download the full model, but only upload 10% of the parameters in each epoch. Finally , in (c), the upload and download is only 10%. 9.2.2 Experiments on A T&T . W e performed similar experiments on the A T&T dataset which consists of faces from 40 dierent people. Initially , we tested the two-participant scenario, where one is the victim, and the other is the adversar y . W e assigned the rst 20 classes to the rst user and the remaining 20 classes to the adversary . An extra class is given to the adversary to inuence the training process. W e ran several congurations with dierent upload rates, see Figure 8. The results show the adversary can get considerably good reconstructions of the targete d face. Some images are noisier (a) θ u = 1 , θ d = 1 (b) θ u = 0 . 1 , θ d = 1 (c) θ u = 0 . 1 , θ d = 0 . 1 Figure 6: Results for the GAN attack on a two-user scenario. Bottom row , samples generated by the GAN. T op row , samples from the training set closest to the ones generate d by the GAN. (a) 100% parameters upload and download. (b) 100% download and 10% upload. (c) 10% upload and download. Figure 7: Collaborative deep learning with 41 participants. All 40 honest users train their respe ctive mo dels on distinct faces. The adversary has no local data. The GAN on the adversary’s device is able to reconstruct the face stored on the victim’s device (ev en when DP is enabled). Original θ u = 1 θ d = 1 θ u = 0 . 1 θ d = 1 θ u = 0 . 1 θ d = 0 . 1 Figure 8: Experimental results on the A T&T Dataset with no DP. Unlike MNIST , images are noisier b ecause this particu- lar dataset is small and the accuracy of the model is signi- cantly aected when upload rates are small. than others, but this can hardly b e improved given that the accuracy of the model tends to stay low for this particular dataset. W e have also implemented a multi-participant scenario, see Fig- ure 7, with 41 participants, 40 of which are honest and 1 is adver- sarial. Each honest participant possesses images pertaining to one class as training data, while the adversary has no training data of his o wn. Namely , the adv ersary only trains on the images produced by the generator (G). The results (with θ u = 1 , θ d = 1 ) are ver y good even when dierential privacy is enabled (Figure 7). 9.3 GAN Attack, No Inuence vs. Inuence on Collaborative Learning One may wonder about the eect of the fake label to the collabora- tive learning. Recall that images generated by the generative model are placed into an articial class to trick the victim into releasing ner details on the targeted class. W e measured the eect of the adversarial inuence, and we experimentally conrmed that its eect is remarkable: The learning gets faster , but also the informa- tion retrieved by the adversary is signicantly better . W e ran the experiments until the accuracy of the mo del on the testing set was above 97% , collaboratively training a CNN model. The datasets of both the adversary and the victim are separated fr om each other , and there are no labels in common. In Figures 9 and 10, we show the result of the passive GAN attack with the standard GAN attack proposed in Section 7, when we are tr ying to recov er , respectively , 0’s and 3’s fr om the rst user . In the top ro w , we sho w the images from the passiv e attack with no inuence and in the bottom row the images from the standard procedure with the inuence of the articial class. The eect of the adversarial inuence is e vident, and images appear much clearer and crisper even after only 50 epochs per participant. During our experiments, we notice d that G starts producing good results as soon as the accuracy of the model reaches 80%. 9.4 GAN Attack on Dierentially Private Collaborative Learning It has been argued in [ 77 ] that dierential privacy can be used to add noise to the parameters of the deep learning model “to ensure that parameter updates do not leak too much information about any individual point in the training dataset. " (Quoted from [ 77 ].) The authors consider only a passive adversary and rely on dierential epoch 5 epoch 20 epo ch 35 epoch 50 epo ch 65 epoch 80 epo ch 95 epoch 110 epoch 125 epoch 140 epoch 155 epoch 5 epoch 20 epo ch 35 epoch 50 epo ch 65 epoch 80 epo ch 95 epoch 110 epoch 125 epoch 140 epoch 155 Figure 9: DCGAN with No inuence vs. inuence in Collab orative Learning for 0 (Zero) epoch 5 epoch 20 epo ch 35 epoch 50 epo ch 65 epoch 80 epo ch 95 epoch 110 epoch 125 epoch 140 epoch 155 epoch 5 epoch 20 epo ch 35 epoch 50 epo ch 65 epoch 80 epo ch 95 epoch 110 epoch 125 epoch 140 epoch 155 Figure 10: DCGAN with No inuence vs. inuence in Collab orative Learning for 3 (Three) Original ϵ c = 100 θ u = 1 ϵ c = 100 θ u = 0 . 1 ϵ c = 10 θ u = 1 ϵ c = 10 θ u = 0 . 1 Figure 11: Experimental results on the A T&T Dataset with 100% download ( ( θ d = 1 ) and DP enable d. Unlike MNIST , images are noisier because this particular dataset is small and the accuracy of the mo del is signicantly aected when upload rates are small. (a) (b) Figure 12: GAN Attack Results on the MNIST Dataset (left: original image, right: generated one) with DP Setup: ϵ c = 0 . 01 , τ = 0 . 0001 , γ = 0 . 001 , θ u = 1 , θ d = 1 . The value of ϵ is so small that the accuracy of the model do es not increase. Since there is no learning, the GAN fails to produce clear results. privacy to mitigate possible leakages that might come from param- eter updates. They highlight two cases of potential leakage: (i) the way how gradient selection is performed and (ii) actual values of (a) (b) Figure 13: GAN Attack Results on the A T&T Dataset (left: original image, right: generated one) with DP Setup: ϵ c = 0 . 01 , τ = 0 . 0001 , γ = 0 . 001 , θ u = 1 , θ d = 1 . The value of ϵ is so small that the accuracy of the model do es not increase. Since there is no learning, the GAN fails to produce clear results. the shared gradients. T o address both of these issues, the approach in [ 77 ] relies on sparse vector technique [ 23 ]. For each epoch (it- eration) of the collaborative learning process, they dene a total (a) ϵ c = 100 , θ u = 1 , θ d = 1 (b) ϵ c = 100 , θ u = 0 . 1 , θ d = 1 (c) ϵ c = 10 , θ u = 1 , θ d = 1 (d) ϵ c = 10 , θ u = 0 . 1 , θ d = 1 Figure 14: Results for the GAN attack on a two-user scenario with Dierential Privacy enabled. Bottom row , samples gen- erated by the GAN. T op row , samples from the training set closest to the ones generated by the GAN. privacy budget ϵ for each participant. This budget is split into c parts, where c is the total number of gradients that can be shared per epoch. A portion of gradients is randomly select such that they are above a threshold ( τ ). They dedicate 8 9 of ϵ c to the selection of the parameters and use the remaining 1 9 to release the value. They rely on the Laplacian mechanism to add noise during selection as well as sharing of the parameters, in agreement with the allocate d privacy budget. T o demonstrate that record-level dierential privacy is ineec- tive against an active adversary , we ran the collaborative learning process between the two participants ( A and V ) with dierential privacy enabled. W e kept the datasets of the participants distinct: In MNIST experiments, V had only records of classes from 0 to 4 and A had records of classes from 5 to 9 plus the articial class that A introduces. For the A T&T experiments, V has records for the rst 20 classes in the dataset and A for the next 20 classes plus the articial class as in Subsection 9.2. During our experiments we kept the download rate ( θ d ) xed at 100%, threshold ( τ ) at 0.0001 and the range ( γ ) at 0.001, similar to [ 77 ]. On Figures 11 and 14, we provide results for a privacy budget per parameter ( ϵ c ) of 100 and 10 and var ying upload rate ( θ u ). Ev en though it takes longer for the models to converge under the dierential privacy constraints, our results demonstrate our claim, i.e., as long as the training process is successful and the model is converging, G can generate goo d results . On the ϵ value. W e observe that the ϵ in [ 77 ] is very large and the eect of dierential privacy may be questionable. Howe ver , with small ϵ , the local mo dels are unable to learn and collaborative learning fails completely . This is consistent with what is reported in [ 77 ]. Indeed, we ran our experiments with tighter privacy con- straints. The generator failed to produce good results but because the local mo del were unable to learn at all. In Figure 12 and 13 we show an example where w e set a tighter privacy b ound, which translates into stronger dierential privacy guarantees, and the GAN is ineective. At the same time, this is expected since the local model and the one in the parameter ser ver ar e unable to learn and collab orative learning is not happening. It is possible to use the techniques in [ 1 ] to bring ϵ down to a single-digit value. How- ever , we stress again that our attack is independent of whatever record-level DP implementation is used. The GAN will generate good samples as long as the discriminator is learning ( see Figure 2). 10 CONCLUSIONS In this work, we propose and implement a novel class of active infer- ence attacks on deep neural networks in a collaborative setting. Our approach relies on Generative Adversarial Networks ( GANs) and is more eective and general than existing information extraction mechanisms. W e b elieve our work will have a signicant impact in the real world as major companies are considering distributed, federated, or decentralize d deep learning approaches to protect the privacy of users. The main point of our research is that collaborative learning is less desirable than the centralized learning approach it is supposed to replace. In collaborative learning, any user may violate the pri- vacy of other users in the system without involving the service provider . Finally , we were not able to devise eective countermeasures against our attack. Solutions may rely on secure multiparty compu- tation or (fully) homomorphic encryption. Howev er: (1) privacy- preserving collaborative learning was introduced as a way to av oid these costly cryptographic primitives [ 77 ], and (2) the solutions we explored based on them would still be susceptible to some forms of our attack. Another approach is to consider dierential privacy at dierent granularities. User or device-level DP would protect against the attacks devised in this paper . However , it’s not clear yet how to build a real system for collaborative learning with de- vice, class, or user-level DP (e.g., users behave and share data in unpredictable ways). Therefore, we leave this subject for future work. A CKNO WLEDGMEN T W e thank Martín Abadi, Matt Fredrikson, Thomas Ristenpart, Vitaly Shmatikov , and Adam Smith for their insightful comments that greatly improv ed our paper . W e are grateful to the authors of [ 77 ] for providing us with the source code of their implementation of privacy-preserving collaborative deep learning. REFERENCES [1] Martín Abadi, Andy Chu, Ian Goodfellow , H Brendan McMahan, Ilya Mironov , Kunal T alwar , and Li Zhang. 2016. Deep learning with dierential privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security . ACM, 308–318. [2] Ahmad Abdulkader , Aparna Lakshmiratan, and Joy Zhang. 2016. Introducing Deep T ext: Facebook’s text understanding engine. (2016). https://tinyurl.com/ jj359dv [3] Martin Arjovsky and Léon Bottou. 2017. T owards principled methods for training generative adversarial networks. In 5th International Conference on Learning Representations (ICLR) . [4] Giuseppe Ateniese, Luigi V Mancini, Angelo Spognardi, Antonio Villani, Domenico Vitali, and Giovanni Felici. 2015. Hacking smart machines with smarter ones: How to extract meaningful data from machine learning clas- siers. International Journal of Security and Networks 10, 3 (2015), 137–150. https://arxiv .org/abs/1306.4447 [5] Gilles Barthe, Noémie Fong, Marco Gaboardi, Benjamin Grégoire, Justin Hsu, and Pierre- Yves Strub . 2016. Advanced probabilistic couplings for dierential privacy . In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security . ACM, 55–67. [6] Y oshua Bengio. 2009. Learning Deep Architectures for AI. Found. Trends Mach. Learn. 2, 1 (Jan. 2009), 1–127. https://doi.org/10.1561/2200000006 [7] Jeremiah Blocki, Anupam Datta, and Joseph Bonneau. 2016. Dierentially Private Password Frequency Lists. In NDSS’16 . [8] Keith Bonawitz, Vladimir Ivanov , Ben Kreuter , Antonio Marcedone, H Bren- dan McMahan, Sarvar Patel, Daniel Ramage, Aaron Segal, and K arn Seth. 2017. Practical Secure Aggregation for Privacy Preserving Machine Learning. (2017). [9] Diane Bouchacourt, Pawan K Mudigonda, and Sebastian Nowozin. 2016. DISCO Nets: DISsimilarity COecients Networks. In Advances in Neural Information Processing Systems . 352–360. [10] Leo Breiman. 2001. Random Forests. Machine Learning 45, 1 (2001), 5–32. [11] Mark Bun and Thomas Steinke. 2016. Concentrated dierential privacy: Simpli- cations, extensions, and lower bounds. In Theory of Cr yptography Conference . Springer , 635–658. [12] Jan Camenisch, Mark Manulis, Gene Tsudik, and Rebecca N. Wright. 2012. Privacy-Oriented Cryptography (Dagstuhl Seminar 12381). Dagstuhl Reports 2 (2012), 165–183. http://drops.dagstuhl.de/opus/volltexte/2013/3755/pdf/dagrep_ v002_i009_p165_s12381.pdf [13] Kamalika Chaudhuri, Claire Monteleoni, and Anand D Sarwate. 2011. Dieren- tially private empirical risk minimization. Journal of machine learning research: JMLR 12 (2011), 1069. [14] Soumith Chintala. 2016. DCGAN.torch: Train your own image generator . (2016). https://github.com/soumith/dcgan.torch [15] Ronan Collobert, Jason W eston, Léon Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa. 2011. Natural language processing (almost) from scratch. Journal of Machine Learning Research 12, Aug (2011), 2493–2537. [16] Angel Alfonso Cruz-Roa, John Edison Arevalo Ovalle, Anant Madabhushi, and Fabio Augusto González Osorio. 2013. A deep learning architecture for image representation, visual interpretability and automated basal-cell car cinoma cancer detection. In International Conference on Medical Image Computing and Computer- Assisted Intervention . Springer Berlin Heidelberg, 403–410. [17] Jerey Dean, Greg Corrado, Rajat Monga, Kai Chen, Matthieu Devin, Mark Mao, Andrew Senior , Paul Tucker , Ke Y ang, Quoc V Le, et al . 2012. Large scale distributed deep networks. In Advances in neural information processing systems . 1223–1231. [18] DeepMind. 2016. DeepMind Health, Clinician-le d. Patient-centred. (2016). https: //deepmind.com/applied/deepmind- health/ [19] Ilias Diakonikolas, Moritz Hardt, and Ludwig Schmidt. 2015. Dierentially private learning of structured discrete distributions. In Advances in Neural Information Processing Systems . 2566–2574. [20] Nathan Dowlin, Ran Gilad-Bachrach, Kim Laine, Kristin Lauter , Naehrig Michael, and John W ernsing. 2016. Cr yptoNets: A pplying Neural Networks to Encr ypted Data with High Throughput and Accuracy . Technical Report MSR-TR-2016-3. http://research.microsoft.com/apps/pubs/default.aspx?id=260989 [21] Cynthia Dwork. 2006. Dierential privacy . In A utomata, Languages and Program- ming, 33rd International Collo quium, ICALP 2006, V enice, Italy , July 10-14, 2006, Proceedings, Part II . Springer Berlin Heidelberg, 1–12. [22] Cynthia Dwork and Moni Naor . 2008. On the diculties of disclosur e prevention in statistical databases or the case for dierential privacy . Journal of Privacy and Condentiality 2, 1 (2008), 8. [23] Cynthia D work and A aron Roth. 2014. The algorithmic foundations of dierential privacy . Foundations and Trends in Theoretical Computer Science 9, 3-4 (2014), 211–407. [24] Cynthia Dwork and Guy N Rothblum. 2016. Concentrated dierential privacy . arXiv preprint arXiv:1603.01887 (2016). [25] Fabienne Eigner , Aniket Kate, Matteo Maei, Francesca Pampaloni, and Ivan Pryvalov. 2014. Dierentially private data aggregation with optimal utility . In Proceedings of the 30th Annual Computer Security Applications Conference . ACM, 316–325. [26] Rasool Fakoor , Faisal Ladhak, Azade Nazi, and Manfred Huber . 2013. Using de ep learning to enhance cancer diagnosis and classication. In The 30th International Conference on Machine Learning (ICML 2013), WHEALTH workshop . [27] Matt Fredrikson, Somesh Jha, and Thomas Ristenpart. 2015. Model inversion attacks that exploit condence information and basic countermeasures. In Pro- ceedings of the 22nd ACM SIGSA C Conference on Computer and Communications Security . ACM, 1322–1333. [28] Matthew Fredrikson, Eric Lantz, Somesh Jha, Simon Lin, David Page, and Thomas Ristenpart. 2014. Privacy in pharmacogenetics: An end-to-end case study of personalized warfarin dosing. In 23rd USENIX Se curity Symposium (USENIX Security 14) . 17–32. [29] Ian Goodfellow , Y oshua Bengio, and Aar on Courville. 2016. Deep learning . MI T Press. [30] Ian Goodfellow , Jean Pouget- Abadie, Mehdi Mirza, Bing Xu, David W arde-Farley , Sherjil Ozair , Aaron Courville, and Y oshua Bengio. 2014. Generative adversarial nets. In Advances in neural information processing systems . 2672–2680. [31] Ian Goodfellow , Jonathon Shlens, and Christian Szegedy . 2015. Explaining and Harnessing Adversarial Examples. In International Conference on Learning Repre- sentations . https://arxiv .org/pdf/1412.6572v3.pdf [32] Ian J Goodfellow . 2014. On distinguishability criteria for estimating generative models. arXiv preprint arXiv:1412.6515 (2014). [33] Google DeepMind. 2016. AlphaGo, the rst computer program to ever b eat a professional player at the game of GO. (2016). https://deepmind.com/alpha- go [34] Alex Graves, Ab del-rahman Mohame d, and Georey Hinton. 2013. Speech recognition with deep recurrent neural networks. In 2013 IEEE international conference on acoustics, speech and signal processing . IEEE, 6645–6649. [35] Andy Greenb erg. 2016. Apple’s ’Dierential Privacy’ Is About Colle cting Y our Data—But Not Y our Data. (2016). https://www .wired.com/2016/06/ apples- dierential- privacy- colle cting- data/ [36] Kathrin Grosse, Nicolas Pap ernot, Praveen Manoharan, Michael Backes, and Patrick McDaniel. 2016. Adversarial Perturbations Against Deep Neural Networks for Malware Classication. arXiv preprint arXiv:1606.04435 (2016). [37] Andreas Haeberlen, Benjamin C. Pierce, and Arjun Narayan. 2011. Dierential Privacy Under Fire. In Proceedings of the 20th USENIX Conference on Security (SEC’11) . USENIX Association, Berkeley , CA, USA, 33–33. http://dl.acm.org/ citation.cfm?id=2028067.2028100 [38] Prateek Jain, Vivek Kulkarni, Abhradeep Thakurta, and Oliver Williams. 2015. T o drop or not to drop: Robustness, consistency and dierential privacy properties of dropout. arXiv:1503.02031 (2015). [39] Shiva Prasad Kasiviswanathan, Homin K Lee, Kobbi Nissim, Sofya Raskhodnikova, and Adam Smith. 2011. What can we learn privately? SIAM J. Comput. 40, 3 (2011), 793–826. [40] Daniel Kifer and Ashwin Machanavajjhala. 2011. No free lunch in data privacy . In Proceedings of the 2011 ACM SIGMOD International Conference on Management of data . ACM, 193–204. [41] Nair Vinod Krizhevsky Alex and Hinton Georey . [n. d.]. CIF AR-10 Dataset. ([n. d.]). https://www.cs.toronto .edu/~kriz/cifar .html [42] Matthew Lai. 2015. Girae: Using deep reinforcement learning to play chess. arXiv preprint arXiv:1509.01549 (2015). [43] Alex M Lamb, Anirudh Goyal ALIAS PARTH GOY AL, Ying Zhang, Saizheng Zhang, Aaron C Courville, and Y oshua Bengio. 2016. Professor forcing: A new algorithm for training recurrent networks. In Advances In Neural Information Processing Systems . 4601–4609. [44] Pavel Laskov et al . 2014. Practical evasion of a learning-based classier: A case study . In Security and Privacy (SP), 2014 IEEE Symposium on . IEEE, 197–211. [45] Y ann LeCun, Y oshua Bengio, and Georey Hinton. 2015. Deep learning. Nature 521, 7553 (2015), 436–444. [46] Y ann LeCun, Corinna Cortes, and Christopher J.C. Burges. 1998. The MNIST database of handwritten digits. (1998). http://yann.lecun.com/exdb/mnist/ [47] Y ann LeCun, Koray Kavukcuoglu, Clément Farabet, et al . 2010. Convolutional networks and applications in vision.. In ISCAS . 253–256. [48] Changchang Liu, Supriyo Chakraborty , and Prateek Mittal. 2016. Dependence Makes Y ou Vulnerable: Dierential Privacy Under Dependent Tuples. In The Network and Distributed System Security Symp osium 2016 (NDSS ’16) . 1322–1333. https://www.internetsociety .org/sites/default/les/blogs- media/ dependence- makes- you- vulnerable- dierential- privacy- under- dep endent- tuples. pdf [49] W arren S. McCulloch and W alter Pitts. 1943. A logical calculus of the ideas immanent in nervous activity . The bulletin of mathematical biophysics 5, 4 (1943), 115–133. https://doi.org/10.1007/BF02478259 [50] Brendan McMahan and Daniel Ramage. 2017. Federated Learning: Collab orative Machine Learning without Centralized Training Data. (2017). https://research. googleblog.com/2017/04/federated- learning- collab orative.html [51] H. Brendan McMahan, Eider Mo ore, Daniel Ramage, and Blaise Ag? era y Ar- cas. 2016. Federated Learning of Deep Networks using Model A veraging. arXiv:1502.01710v5 (2016). [52] Richard McPherson, Reza Shokri, and Vitaly Shmatikov . 2016. Defeating Image Obfuscation with Deep Learning. arXiv:1609.00408 (2016). [53] Frank McSherry . 2016. Dierential Privacy and Correlated Data. (2016). https: //github.com/frankmcsherry/blog/blob/master/posts/2016- 08- 29.md [54] Frank McSherry . 2016. Lunchtime for Data Privacy . (2016). https://github.com/ frankmcsherry/blog/blob/master/posts/2016- 08- 16.md [55] Frank McSherr y and Ilya Mironov . 2009. Dierentially private recommender systems: building privacy into the Netix Prize contenders. In Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining . ACM, 627–636. [56] Lars Mescheder , Sebastian Nowozin, and Andreas Geiger . 2017. Adversarial V ariational Bayes: Unifying Variational Autoencoders and Generative Adversarial Networks. arXiv preprint arXiv:1701.04722 (2017). [57] Cade Metz. 2016. Google’s GO victory is just a glimpse of how powerful ai will be. (2016). https://tinyurl.com/l6ddhg9 [58] Prateek Mittal. 2016. Dierential Privacy is Vulnerable to Correlated Data Introducing Dependent Dierential Privacy . (2016). https://tinyurl.com/l3lx7qh [59] V olo dymyr Mnih, Adria Puigdomenech Badia, Mehdi Mirza, Alex Graves, Timo- thy P Lillicrap, Tim Harley , David Silver , and Koray K avukcuoglu. 2016. Asyn- chronous methods for deep reinforcement learning. arXiv:1602.01783 (2016). [60] V olo dymyr Mnih, Koray Kavukcuoglu, David Silver , Alex Graves, Ioannis Antonoglou, Daan Wierstra, and Martin Riedmiller . 2013. P laying atari with deep reinforcement learning. arXiv:1312.5602 (2013). [61] Payman Mohassel and Y upeng Zhang. 2017. SecureML: A System for Scalable Privacy-Preserving Machine Learning. In IEEE Symposium on Security and Pri- vacy . [62] Arjun Narayan, Ariel Feldman, Antonis Papadimitriou, and Andreas Haeberlen. 2015. V eriable dierential privacy . In Proceedings of the Tenth European Confer- ence on Computer Systems . ACM, 28. [63] Olga Ohrimenko, Felix Schuster , Cédric Fournet, Aastha Mehta, Sebastian Nowozin, Kapil Vaswani, and Manuel Costa. 2016. Oblivious Multi-Party Machine Learning on Trusted Processors. In USENIX Se curity . [64] Aaron van den Oord, Sander Dieleman, Heiga Zen, Karen Simonyan, Oriol Vinyals, Alex Graves, Nal Kalchbrenner , Andrew Senior , and Koray Kavukcuoglu. 2016. W aveNet: A generative model for raw audio. arXiv:1609.03499 (2016). [65] Nicolas Papernot, Patrick McDaniel, and Ian Goodfellow . 2016. Transferability in Machine Learning: from Phenomena to Black-Box Attacks using Adversarial Samples. arXiv preprint arXiv:1605.07277 (2016). [66] Nicolas Pap ernot, Patrick McDaniel, Somesh Jha, Matt Fredrikson, Z Berkay Celik, and Ananthram Swami. 2015. The Limitations of Deep Learning in Adversarial Settings. Proce edings of the 1st IEEE European Symposium on Security and Privacy (2015). [67] Manas Pathak, Shantanu Rane, and Bhiksha Raj. 2010. Multiparty dierential privacy via aggregation of locally trained classiers. In Advances in Neural Infor- mation Processing Systems . 1876–1884. [68] Guim Perarnau, Joost van de W eijer , Bogdan Raducanu, and Jose M Álvarez. 2016. Invertible Conditional GANs for image editing. arXiv preprint (2016). [69] NhatHai Phan, Yue W ang, Xintao Wu, and Dejing Dou. 2016. Dierential Pri- vacy Preservation for Deep Auto-Encoders: an Application of Human Behavior Prediction. In Proceedings of the 30th AAAI Conference on A rticial Intelligence, AAAI . 12–17. [70] Alec Radford, Luke Metz, and Soumith Chintala. 2016. Unsupervised Representa- tion Learning with Deep Convolutional Generative Adversarial Networks. In 4th International Conference on Learning Representations . [71] C. E. Rasmussen and C. K. I. Williams. 2006. Gaussian Processes for Machine Learning . MI T Press, Cambridge, MA. [72] Tim Salimans, Ian Goodfellow , W ojciech Zaremba, Vicki Cheung, Alec Radford, and Xi Chen. 2016. Improved techniques for training gans. In A dvances in Neural Information Processing Systems . 2226–2234. [73] Ferdinando S Samaria and Andy C Harter . 1994. Parameterisation of a stochastic model for human face identication. In Applications of Computer Vision, 1 994., Proceedings of the Second IEEE Workshop on . IEEE, 138–142. [74] Anand D Sar wate and Kamalika Chaudhuri. 2013. Signal processing and machine learning with dierential privacy: Algorithms and challenges for continuous data. IEEE signal processing magazine 30, 5 (2013), 86–94. [75] Jürgen Schmidhuber . 2015. Deep learning in neural networks: An overview . Neural networks 61 (2015), 85–117. [76] Bernhard Scholkopf and Alexander J Smola. 2001. Learning with kernels: support vector machines, regularization, optimization, and beyond . MI T press. [77] Reza Shokri and Vitaly Shmatikov. 2015. Privacy-Preserving Deep Learning. In Proceedings of the 22Nd ACM SIGSAC Conference on Computer and Commu- nications Security (CCS ’15) . ACM, 1310–1321. https://doi.org/10.1145/2810103. 2813687 [78] Reza Shokri, Marco Stronati, Congzheng Song, and Vitaly Shmatikov . 2017. Mem- bership Inference Attacks against Machine Learning Models. In IEEE Symposium on Security and Privacy (S&P), Oakland . [79] Shuang Song, Kamalika Chaudhuri, and Anand D Sarwate. 2013. Stochastic gradient descent with dierentially private up dates. In Global Conference on Signal and Information Processing (GlobalSIP), 2013 IEEE . IEEE, 245–248. [80] Praveen De epak Srinivasan, Ror y Fearon, Cagdas Alcicek, Arun Sarath Nair , Samuel Blackwell, V edav yas Panneershelvam, Alessandro De Maria, V olodymyr Mnih, Koray Kavukcuoglu, David Silver , et al . 2016. Distributed training of reinforcement learning systems. (Feb. 4 2016). US Patent App. 15/016,173. [81] Christian Szege dy , W ojciech Zaremba, Ilya Sutskever , Joan Bruna, Dumitru Erhan, Ian Goodfellow , and Rob Fergus. 2014. Intriguing properties of neural networks. In International Conference on Learning Representations . http://ar xiv .org/abs/1312. 6199 [82] Y aniv T aigman, Ming Y ang, Marc’ A urelio Ranzato, and Lior W olf. 2014. De epFace: Closing the Gap to Human-Level Performance in Face V erication. In Pr oceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR ’14) . IEEE Computer Society , W ashington, DC, USA, 1701–1708. https://doi.org/ 10.1109/CVPR.2014.220 [83] Florian Tramèr , Fan Zhang, Ari Juels, Michael K Reiter, and Thomas Ristenpart. 2016. Stealing Machine Learning Models via Prediction APIs. In USENIX Security . [84] Vladimir Naumovich Vapnik and Vlamimir V apnik. 1998. Statistical learning theory . V ol. 1. Wiley New Y ork. [85] Martin J W ainwright, Michael I Jordan, and John C Duchi. 2012. Privacy aware learning. In Advances in Neural Information Processing Systems . 1430–1438. [86] Pengtao Xie, Misha Bilenko, T om Finley , Ran Gilad-Bachrach, Kristin Lauter , and Michael Naehrig. 2014. Cr ypto-nets: Neural networks ov er encrypted data. arXiv preprint arXiv:1412.6181 (2014). [87] W eilin Xu, Yanjun Qi, and David Evans. 2016. Automatically evading classiers. In NDSS’16 . [88] Jun Zhang, Zhenjie Zhang, Xiaokui Xiao, Yin Y ang, and Marianne Winslett. 2012. Functional mechanism: regression analysis under dierential privacy. Proceedings of the VLDB Endowment 5, 11 (2012), 1364–1375. [89] T ong Zhang. 2004. Solving large scale linear prediction problems using stochas- tic gradient descent algorithms. In Proceedings of the twenty-rst international conference on Machine learning . ACM, 116. [90] Xiang Zhang and Y ann André LeCun. 2016. T ext Understanding from Scratch. arXiv preprint arXiv:1502.01710v5 (2016). [91] Martin Zinkevich, Markus W eimer, Lihong Li, and Alex J Smola. 2010. Parallelized stochastic gradient descent. In Advances in neural information processing systems . 2595–2603. A SYSTEM ARCHI TECT URE Figure 15: Convolutional Neural Network Architecture use d for MNIST related experiments, as printed by T orch. Note that the same architecture is use d for both the collaboratively trained mo del and the local discriminator (D) model used by the Adversary Figure 16: Generator Mo del Architecture used in MNIST experiments Figure 17: Architecture of the Collaborative Model and the Discriminator (D) utilized in A T&T Dataset related experiments Figure 18: Generator (G) Architecture used in A T&T Dataset related experiments, as printed by T orch7
Original Paper
Loading high-quality paper...
Comments & Academic Discussion
Loading comments...
Leave a Comment