<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" ><generator uri="https://jekyllrb.com/" version="3.10.0">Jekyll</generator><link href="https://qdata.github.io//qdata-page/feed.xml" rel="self" type="application/atom+xml" /><link href="https://qdata.github.io//qdata-page/" rel="alternate" type="text/html" /><updated>2025-12-06T16:34:29-05:00</updated><id>https://qdata.github.io//qdata-page/feed.xml</id><title type="html">Qdata Team Research Blog Site</title><subtitle>Research Blogs of Qdata Team in the University of Virignia Computer Science Department.</subtitle><author><name>Qdata</name></author><entry><title type="html">Zhe’s PhD Defense - Toward Out-Of-Distribution Generalization Of Deep Learning Models</title><link href="https://qdata.github.io//qdata-page/aiself/zhe-defense/" rel="alternate" type="text/html" title="Zhe’s PhD Defense - Toward Out-Of-Distribution Generalization Of Deep Learning Models" /><published>2024-04-02T00:00:00-04:00</published><updated>2024-04-02T00:00:00-04:00</updated><id>https://qdata.github.io//qdata-page/aiself/zhe-defense</id><content type="html" xml:base="https://qdata.github.io//qdata-page/aiself/zhe-defense/"><![CDATA[<h2 id="phd-dissertation-defense-by-zhe-wang">Ph.D. Dissertation Defense by Zhe Wang,</h2>
<ul>
  <li>Tues., 04/02/24, at 12:00PM (ET)</li>
</ul>

<h4 id="committee">Committee:</h4>

<p>Matthew Dwyer, Committee Chair  (CS/SEAS/UVA)
Yanjun Qi, Advisor (CS/SEAS, SDS, SM/UVA)
Miaomiao Zhang (ECE/CS/SEAS/UVA)
Jianhui Zhou (Statistics/College/UVA)
Vicente Ordonez (CS/Rice University)</p>

<h4 id="title-toward-out-of-distribution-generalization-of-deep-learning-models">Title: Toward Out-Of-Distribution Generalization Of Deep Learning Models</h4>

<ul>
  <li>Abstract : Deep learning models, especially deep neural networks (DNNs), perform extremely well when the testing and training distributions align. However, real-world scenarios often witness shifts in data distribution across domains, and tasks, over time, and are influenced by adversarial attacks. Such shifts from the training to testing distribution present challenges, resulting in performance degradation of DNNs. The varied testing distributions from diverse users underscore the urgent necessity to understand OOD problems and design methods to mitigate OOD generalization challenges. Therefore, this dissertation develops methods and strategies to enhance DNNs’ ability to generalize to unseen distributions. First, we focus on generalizing DNNs to unknown domains, in which no prior information about testing domains is available during training. We propose a novel optimization approach that learns principal gradients from eigenvectors of training optimization trajectories. This robust gradient design forces the training to ignore domain-dependent noise signals and updates all training domains with a robust direction covering the main components of parameter dynamics. Second, we focus on designing strategies to generalize DNNs to unseen tasks (i.e. meta-learning), for instance, a new unknown RL task with few demonstration trajectories. The main challenge is to infer the potential identity of a new task from a limited number of annotated samples. We propose modeling a new task’s identity as a stochastic variable and encoding it with a stochastic neural network. This task identity design helps meta-learning to adapt shared training knowledge to a new current task. When solving similar task generalization issues in offline RL, we further propose learning from the RL transition dynamic and reward function to capture a task’s identity. Third, deep learning models should not only perform well on clean, legitimate data distribution but also on data that has been subjected to adversarial attacks. Entering the era of large foundation models, we focus on techniques to craft adversarial attackers for jailbreaking pretrained large language models (LLMs) due to their prevalent recent adoptions. We design a new objective, which learns adversarial suffixes with much cheaper queries and higher attack success rate. The learned suffixes also demonstrate higher transferability across LLMs. In the thesis, we validate the effectiveness of our methods across image classification and completion, wealth index regression from satellite images, robotic control, real-world temperature forecasting, and natural language generation.</li>
</ul>]]></content><author><name>Qdata</name></author><category term="AIself" /><category term="Tutorials" /><category term="Graph-discriminative" /><category term="Semi-SelfLabel" /><category term="RL-Generalization" /><category term="multi-transfer" /><summary type="html"><![CDATA[Ph.D. Dissertation Defense by Zhe Wang, Tues., 04/02/24, at 12:00PM (ET) Committee:]]></summary></entry><entry><title type="html">Arsh’s PhD Defense - Relational Structure Discovery for Deep Learning</title><link href="https://qdata.github.io//qdata-page/aiself/Arsh-defense/" rel="alternate" type="text/html" title="Arsh’s PhD Defense - Relational Structure Discovery for Deep Learning" /><published>2022-06-29T00:00:00-04:00</published><updated>2022-06-29T00:00:00-04:00</updated><id>https://qdata.github.io//qdata-page/aiself/Arsh-defense</id><content type="html" xml:base="https://qdata.github.io//qdata-page/aiself/Arsh-defense/"><![CDATA[<h2 id="arshdeep-sekhons-phd-defense">Arshdeep Sekhon’s PhD Defense</h2>
<ul>
  <li>June 29, 2022.</li>
</ul>

<h4 id="committee">Committee:</h4>
<p>Yanjun Qi (Advisor)(CS/SEAS/UVA)
Matthew Dwyer (CS/SEAS/UVA)
Yangfeng Ji (CS/SEAS/UVA)
Vicente Ordonez (CS/SEAS/UVA)
Jianhui Zhou (Department of Statistics/UVA)</p>

<h4 id="title-relational-structure-discovery-for-deep-learning">Title: Relational Structure Discovery for Deep Learning</h4>

<ul>
  <li>Abstract: 
Graph structure is ubiquitous: from physical relationships to biological interactions to social networks, and many more spread across the universe. Not only is the world around us rich in relational structure, but our mental model of the world is also structured: we think, reason, and communicate in terms of entities and their relations. Such a graph-structured real world calls for artificial intelligence methods that think like humans and hence employ this structure for decision making. Realizing such a framework requires known structure/graph and models that can ingest these non-linear graphical inputs. In cases of a latent unknown graph structure, state-of-the-art deep learning models either focus on task-agnostic statistical dependency learning or diverge from explicit feature dependencies during prediction. We bridge this gap and introduce methods for jointly learning and incorporating graph-based relational knowledge into state-of-the-art deep learning models to help improve (1) predictions, (2) interpretability, (3) post-hoc interpretations, and (4) test datasets selection. Specifically, we contribute methods that enable learning graphical relationships from data without such a ground truth graph. Furthermore, we introduce plug-and-play methods that bias deep learning models to include the learned graph explicitly for improving the aforementioned downstream tasks. We demonstrate our methods’ capabilities on simulated, tabular, NLP, and vision tasks.</li>
</ul>]]></content><author><name>Qdata</name></author><category term="AIself" /><category term="Tutorials" /><category term="Graph-discriminative" /><category term="Graph-generative" /><category term="multi-Graph-generative" /><summary type="html"><![CDATA[Arshdeep Sekhon’s PhD Defense June 29, 2022.]]></summary></entry><entry><title type="html">JackL’s PhD Defense - Modeling interactions with Deep Learning</title><link href="https://qdata.github.io//qdata-page/aiself/JackL-defense/" rel="alternate" type="text/html" title="JackL’s PhD Defense - Modeling interactions with Deep Learning" /><published>2021-07-20T00:00:00-04:00</published><updated>2021-07-20T00:00:00-04:00</updated><id>https://qdata.github.io//qdata-page/aiself/JackL-defense</id><content type="html" xml:base="https://qdata.github.io//qdata-page/aiself/JackL-defense/"><![CDATA[<h2 id="phd-dissertation-defense-by-jack-lanchantin">Ph.D. Dissertation Defense by Jack Lanchantin</h2>
<ul>
  <li>Tuesday, July 20th, 2021 at 2:00 PM (ET), via Zoom.</li>
</ul>

<h4 id="committee">Committee:</h4>
<ul>
  <li>Vicente Ordóñez Román, Committee Chair, (CS/SEAS/UVA)</li>
  <li>Yanjun Qi, advisor, (CS/SEAS/UVA)</li>
  <li>Yangfeng Ji (CS/SEAS/UVA)</li>
  <li>Clint Miller (Public Health Sciences/SOM/UVA)</li>
  <li>Casey Greene (Biochemistry &amp; Molecular Genetics/SOM/University of Colorado)</li>
</ul>

<h3 id="title-modeling-interactions-with-deep-learning">Title: Modeling interactions with Deep Learning</h3>

<ul>
  <li>Abstract: Interacting systems are highly prevalent in many real-world settings, including genomics, proteomics, and images. The dynamics of complex systems are often explained as a composition of entities and their interaction graphs. In this dissertation, we design state-of-the-art deep neural networks for interaction-oriented representation learning. Learning such structure representations from data can provide semantic clarity, ease of reasoning for generating new knowledge, and potentially causal interpretation. We consider three different types of interactions: 1) interactions within a particular input sample, 2) interactions between multiple input samples, and 3) interactions between output labels. For each type of interaction, we design novel models to tackle a real-world problem and validate our results both quantitatively and visually.</li>
</ul>]]></content><author><name>Qdata</name></author><category term="AIself" /><category term="EpiGenome" /><category term="DeepSequence" /><category term="6-Interpret" /><category term="Tutorials" /><category term="Graph-discriminative" /><category term="Semi-SelfLabel" /><summary type="html"><![CDATA[Ph.D. Dissertation Defense by Jack Lanchantin Tuesday, July 20th, 2021 at 2:00 PM (ET), via Zoom. Committee: Vicente Ordóñez Román, Committee Chair, (CS/SEAS/UVA) Yanjun Qi, advisor, (CS/SEAS/UVA) Yangfeng Ji (CS/SEAS/UVA) Clint Miller (Public Health Sciences/SOM/UVA) Casey Greene (Biochemistry &amp; Molecular Genetics/SOM/University of Colorado) Title: Modeling interactions with Deep Learning]]></summary></entry><entry><title type="html">DrQi’s tutorial talk on “Make Deep Learning Interpretable for Sequential Data Analysis in Biomedicine” (Including our work on DeepChrome - AttentiveChrome - GCNChrome - DeepMotif - DeepVHPPI - MotifTransformer)</title><link href="https://qdata.github.io//qdata-page/aibiomed/deep-bioSeq-tutorial/" rel="alternate" type="text/html" title="DrQi’s tutorial talk on “Make Deep Learning Interpretable for Sequential Data Analysis in Biomedicine” (Including our work on DeepChrome - AttentiveChrome - GCNChrome - DeepMotif - DeepVHPPI - MotifTransformer)" /><published>2021-07-01T00:00:00-04:00</published><updated>2021-07-01T00:00:00-04:00</updated><id>https://qdata.github.io//qdata-page/aibiomed/deep-bioSeq-tutorial</id><content type="html" xml:base="https://qdata.github.io//qdata-page/aibiomed/deep-bioSeq-tutorial/"><![CDATA[<p>I gave a tutorial talk  at UVA-VADC Seminar Series 2021 and at monthly NIH Data Science Showcase seminar.</p>

<h1 id="title-make-deep-learning-interpretable-for-sequential-data-analysis-in-biomedicine">Title: Make Deep Learning Interpretable for Sequential Data Analysis in Biomedicine</h1>

<h2 id="slide-pdf"><a href="/qdata-page//talk/2021620-VADC-deepGenomics.pdf">Slide PDF</a></h2>

<hr />

<p>This tutorial includes four of our recent papers:</p>

<h2 id="tool-deepchrome-deep-learning-for-predicting-gene-expression-from-histone-modifications">Tool DeepChrome: deep-learning for predicting gene expression from histone modifications</h2>

<ul>
  <li>
    <p>Paper:  <a href="https://academic.oup.com/bioinformatics/article-abstract/32/17/i639/2450757/DeepChrome-deep-learning-for-predicting-gene">@Bioinformatics</a></p>
  </li>
  <li>
    <p><a href="https://github.com/QData/DeepChrome">GitHub</a></p>
  </li>
</ul>

<h3 id="tool-attentivechrome-attend-and-predict-using-deep-attention-model-to-understand-gene-regulation-by-selective-attention-on-chromatin">Tool AttentiveChrome: Attend and Predict: Using Deep Attention Model to Understand Gene Regulation by Selective Attention on Chromatin</h3>

<ul>
  <li>
    <p>Paper:  Published at <a href="https://papers.nips.cc/paper/7255-attend-and-predict-understanding-gene-regulation-by-selective-attention-on-chromatin.pdf">NeurIPS2017</a></p>
  </li>
  <li>
    <p><a href="https://github.com/QData/AttentiveChrome">GitHub</a></p>
  </li>
</ul>

<h3 id="tool-gcnchrome-graph-convolutional-networks-for-epigenetic-state-prediction-using-both-sequence-and-3d-genome-data">Tool: GCNChrome: Graph Convolutional Networks for Epigenetic State Prediction Using Both Sequence and 3D Genome Data</h3>

<ul>
  <li>
    <p>Paper @<a href="https://academic.oup.com/bioinformatics/article/36/Supplement_2/i659/6055904">Bioinformatics</a></p>
  </li>
  <li>
    <p>GitHub: <a href="https://github.com/QData/ChromeGCN">https://github.com/QData/ChromeGCN</a></p>
  </li>
</ul>

<h3 id="tool-transfer-learning-for-predicting-virus-host-protein-interactions-for-novel-virus-sequences">Tool: Transfer Learning for Predicting Virus-Host Protein Interactions for Novel Virus Sequences</h3>

<ul>
  <li>
    <p>PDF @ <a href="https://www.biorxiv.org/content/10.1101/2020.12.14.422772v2">ACM BCB21</a></p>
  </li>
  <li>
    <p>GitHub <a href="https://github.com/QData/DeepVHPPI">https://github.com/QData/DeepVHPPI</a></p>
  </li>
</ul>

<p><br /></p>

<p>Thanks for reading!</p>]]></content><author><name>Qdata</name></author><category term="AIbiomed" /><category term="Tutorials" /><category term="6-Interpret" /><category term="Protein" /><category term="DeepSequence" /><category term="BioGraph" /><category term="ToolKit" /><summary type="html"><![CDATA[I gave a tutorial talk at UVA-VADC Seminar Series 2021 and at monthly NIH Data Science Showcase seminar.]]></summary></entry><entry><title type="html">ACM BCB - Transfer Learning for Predicting Virus-Host Protein Interactions for Novel Virus Sequences</title><link href="https://qdata.github.io//qdata-page/aibiomed/aiself/TransformerPPI/" rel="alternate" type="text/html" title="ACM BCB - Transfer Learning for Predicting Virus-Host Protein Interactions for Novel Virus Sequences" /><published>2021-06-11T00:00:00-04:00</published><updated>2021-06-11T00:00:00-04:00</updated><id>https://qdata.github.io//qdata-page/aibiomed/aiself/TransformerPPI</id><content type="html" xml:base="https://qdata.github.io//qdata-page/aibiomed/aiself/TransformerPPI/"><![CDATA[<p><a name="semippi"></a></p>

<h3 id="title-transfer-learning-for-predicting-virus-host-protein-interactions-for-novel-virus-sequences">Title: Transfer Learning for Predicting Virus-Host Protein Interactions for Novel Virus Sequences</h3>

<ul>
  <li>authors: Jack Lanchantin, Tom Weingarten, Arshdeep Sekhon, Clint Miller, Yanjun Qi</li>
  <li>2021 ACM Conference on Bioinformatics, Computational Biology, and Health Informatics (ACM BCB)</li>
</ul>

<h3 id="pdf--bioarxiv">PDF @ <a href="https://www.biorxiv.org/content/10.1101/2020.12.14.422772v2">BioArxiv</a></h3>

<ul>
  <li>GitHub <a href="https://github.com/QData/DeepVHPPI">https://github.com/QData/DeepVHPPI</a></li>
</ul>

<h3 id="talk-slide">Talk: <a href="/qdata-page//pic/semiPPI.pdf">Slide</a></h3>

<h3 id="abstract">Abstract</h3>

<p>Viruses such as SARS-CoV-2 infect the human body by forming interactions between virus proteins and human proteins. However, experimental methods to find protein interactions are inadequate: large scale experiments are noisy, and small scale experiments are slow and expensive. Inspired by the recent successes of deep neural networks, we hypothesize that deep learning methods are well-positioned to aid and augment biological experiments, hoping to help identify more accurate virus-host protein interaction maps. Moreover, computational methods can quickly adapt to predict how virus mutations change protein interactions with the host proteins.</p>

<p>We propose DeepVHPPI, a novel deep learning framework combining a self-attention-based transformer architecture and a transfer learning training strategy to predict interactions between human proteins and virus proteins that have novel sequence patterns. We show that our approach outperforms the state-of-the-art methods significantly in predicting Virus–Human protein interactions for SARS-CoV-2, H1N1, and Ebola. In addition, we demonstrate how our framework can be used to predict and interpret the interactions of mutated SARS-CoV-2 Spike protein sequences.</p>

<p>We make all of our data and code available on GitHub <a href="https://github.com/QData/DeepVHPPI">https://github.com/QData/DeepVHPPI</a>.</p>

<p><img src="/qdata-page//pic/deepVH2.png" alt="demo1" class="img-responsive" />
<img src="/qdata-page//pic/deepVH3.png" alt="demo1" class="img-responsive" />
<img src="/qdata-page//pic/deepVH4.png" alt="demo1" class="img-responsive" /></p>

<h3 id="citations">Citations</h3>

<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre class="highlight"><code>@article {Lanchantin2020.12.14.422772,
	author = {Lanchantin, Jack and Weingarten, Tom and Sekhon, Arshdeep and Miller, Clint and Qi, Yanjun},
	title = {Transfer Learning for Predicting Virus-Host Protein Interactions for Novel Virus Sequences},
	elocation-id = {2020.12.14.422772},
	year = {2021},
	doi = {10.1101/2020.12.14.422772},
	publisher = {Cold Spring Harbor Laboratory},
	abstract = {Viruses such as SARS-CoV-2 infect the human body by forming interactions between virus proteins and human proteins. However, experimental methods to find protein interactions are inadequate: large scale experiments are noisy, and small scale experiments are slow and expensive. Inspired by the recent successes of deep neural networks, we hypothesize that deep learning methods are well-positioned to aid and augment biological experiments, hoping to help identify more accurate virus-host protein interaction maps. Moreover, computational methods can quickly adapt to predict how virus mutations change protein interactions with the host proteins.We propose DeepVHPPI, a novel deep learning framework combining a self-attention-based transformer architecture and a transfer learning training strategy to predict interactions between human proteins and virus proteins that have novel sequence patterns. We show that our approach outperforms the state-of-the-art methods significantly in predicting Virus{\textendash}Human protein interactions for SARS-CoV-2, H1N1, and Ebola. In addition, we demonstrate how our framework can be used to predict and interpret the interactions of mutated SARS-CoV-2 Spike protein sequences.Availability We make all of our data and code available on GitHub https://github.com/QData/DeepVHPPI.ACM Reference Format Jack Lanchantin, Tom Weingarten, Arshdeep Sekhon, Clint Miller, and Yanjun Qi. 2021. Transfer Learning for Predicting Virus-Host Protein Interactions for Novel Virus Sequences. In Proceedings of ACM Conference (ACM-BCB). ACM, New York, NY, USA, 10 pages. https://doi.org/??Competing Interest StatementThe authors have declared no competing interest.},
	URL = {https://www.biorxiv.org/content/early/2021/06/08/2020.12.14.422772},
	eprint = {https://www.biorxiv.org/content/early/2021/06/08/2020.12.14.422772.full.pdf},
	journal = {bioRxiv}
}

</code></pre></div></div>

<h3 id="support-or-contact">Support or Contact</h3>

<p>Having trouble with our tools? Please <a href="mailto:jacklanchantin@gmail.com">contact Jack</a> and we’ll help you sort it out.</p>]]></content><author><name>Qdata</name></author><category term="AIbiomed" /><category term="AIself" /><category term="Protein" /><category term="DeepSequence" /><category term="Semi-SelfLabel" /><category term="Predict-edges" /><category term="BioGraph" /><category term="Represent-ngram" /><category term="multi-transfer" /><summary type="html"><![CDATA[]]></summary></entry><entry><title type="html">Dr Qi’s Invited Talks on textattack</title><link href="https://qdata.github.io//qdata-page/aitrust/textAttackHMI/" rel="alternate" type="text/html" title="Dr Qi’s Invited Talks on textattack" /><published>2021-04-13T00:00:00-04:00</published><updated>2021-04-13T00:00:00-04:00</updated><id>https://qdata.github.io//qdata-page/aitrust/textAttackHMI</id><content type="html" xml:base="https://qdata.github.io//qdata-page/aitrust/textAttackHMI/"><![CDATA[<h3 id="on-june-24th-2021-i-gave-an-invited-talk-at-the-science-academy-machine-learning-summer-school-on-textattack-generalizing-adversarial-examples-to">On June 24th, 2021, I gave an invited talk at the Science Academy Machine Learning Summer School on “TextAttack: Generalizing Adversarial Examples to</h3>
<p>Natural Language Processing”</p>

<ul>
  <li><a href="/qdata-page/pic/20210625-TML-textAttack-v3-main.pdf">TalkSlide</a></li>
</ul>

<h3 id="previous-version-of-the-tutorial-on-april-14-2021-i-gave-an-invited-talk-at-the-uva-human-and-machine-intelligence-seminar">Previous version of the tutorial: On April 14 2021, I gave an invited talk at the UVA Human and Machine Intelligence Seminar:</h3>

<ul>
  <li><a href="/qdata-page/pic/20210414-HMI-textAttack.pdf">TalkSlide</a></li>
</ul>]]></content><author><name>Qdata</name></author><category term="AItrust" /><category term="1-Evasion" /><category term="ToolKit" /><category term="Tutorials" /><category term="textattack" /><summary type="html"><![CDATA[On June 24th, 2021, I gave an invited talk at the Science Academy Machine Learning Summer School on “TextAttack: Generalizing Adversarial Examples to Natural Language Processing”]]></summary></entry><entry><title type="html">CVPR - General Multi-label Image Classification with Transformers</title><link href="https://qdata.github.io//qdata-page/aifastconnectome/aiself/Ctrans/" rel="alternate" type="text/html" title="CVPR - General Multi-label Image Classification with Transformers" /><published>2021-03-01T00:00:00-05:00</published><updated>2021-03-01T00:00:00-05:00</updated><id>https://qdata.github.io//qdata-page/aifastconnectome/aiself/Ctrans</id><content type="html" xml:base="https://qdata.github.io//qdata-page/aifastconnectome/aiself/Ctrans/"><![CDATA[<h3 id="title-general-multi-label-image-classification-with-transformers">Title: General Multi-label Image Classification with Transformers</h3>

<h3 id="paper-arxivversion">Paper <a href="https://arxiv.org/abs/2011.14027">ArxivVersion</a></h3>

<h3 id="github-httpsgithubcomqdatac-tran">GitHub: <a href="https://github.com/QData/C-Tran">https://github.com/QData/C-Tran</a></h3>

<h3 id="abstract">Abstract</h3>
<p>Multi-label image classification is the task of predicting a set of labels corresponding to objects, attributes or other entities present in an image. In this work we propose the Classification Transformer (C-Tran), a general framework for multi-label image classification that leverages Transformers to exploit the complex dependencies among visual features and labels. Our approach consists of a Transformer encoder trained to predict a set of target labels given an input set of masked labels, and visual features from a convolutional neural network. A key ingredient of our method is a label mask training objective that uses a ternary encoding scheme to represent the state of the labels as positive, negative, or unknown during training. Our model shows state-of-the-art performance on challenging datasets such as COCO and Visual Genome. Moreover, because our model explicitly represents the uncertainty of labels during training, it is more general by allowing us to produce improved results for images with partial or extra label annotations during inference. We demonstrate this additional capability in the COCO, Visual Genome, News500, and CUB image datasets.</p>

<h3 id="citations">Citations</h3>

<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre class="highlight"><code>@article{lanchantin2020general,
      title={General Multi-label Image Classification with Transformers}, 
      author={Jack Lanchantin and Tianlu Wang and Vicente Ordonez and Yanjun Qi},
      year={2020},
      eprint={2011.14027},
      archivePrefix={arXiv, CVPR2021},
      primaryClass={cs.CV}
}
</code></pre></div></div>

<h3 id="support-or-contact">Support or Contact</h3>

<p>Having trouble with our tools? Please <a href="mailto:jjl5sw@virginia.edu">contact Jack Lanchantin</a> and we’ll help you sort it out.</p>]]></content><author><name>Qdata</name></author><category term="AIfastConnectome" /><category term="AIself" /><category term="Graph-discriminative" /><category term="Semi-SelfLabel" /><category term="ToolKit" /><summary type="html"><![CDATA[Title: General Multi-label Image Classification with Transformers]]></summary></entry><entry><title type="html">AAAI - Curriculum Labeling- Self-paced Pseudo-Labeling for Semi-Supervised Learning</title><link href="https://qdata.github.io//qdata-page/aiself/pseudolabel/" rel="alternate" type="text/html" title="AAAI - Curriculum Labeling- Self-paced Pseudo-Labeling for Semi-Supervised Learning" /><published>2021-01-11T00:00:00-05:00</published><updated>2021-01-11T00:00:00-05:00</updated><id>https://qdata.github.io//qdata-page/aiself/pseudolabel</id><content type="html" xml:base="https://qdata.github.io//qdata-page/aiself/pseudolabel/"><![CDATA[<h3 id="title-curriculum-labeling--self-paced-pseudo-labeling-for-semi-supervised-learning">Title: Curriculum Labeling- Self-paced Pseudo-Labeling for Semi-Supervised Learning”</h3>

<h4 id="at-the-thirty-fifth-aaai-conference-on-artificial-intelligence-aaai-21-acceptance-rate-21">at the Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI-21) (acceptance rate: 21%))</h4>

<h4 id="authors-paola-cascante-bonilla-fuwen-tan-yanjun-qi-vicente-ordonez">authors: Paola Cascante-Bonilla, Fuwen Tan, Yanjun Qi, Vicente Ordonez</h4>

<h3 id="paper-arxiv">Paper <a href="https://arxiv.org/abs/2001.06001">Arxiv</a></h3>

<h3 id="abstract">Abstract</h3>
<p>In this paper we revisit the idea of pseudo-labeling in the context of semi-supervised learning where a learning algorithm has access to a small set of labeled samples and a large set of unlabeled samples. Pseudo-labeling works by applying pseudo-labels to samples in the unlabeled set by using a model trained on the combination of the labeled samples and any previously pseudo-labeled samples, and iteratively repeating this process in a self-training cycle. Current methods seem to have abandoned this approach in favor of consistency regularization methods that train models under a combination of different styles of self-supervised losses on the unlabeled samples and standard supervised losses on the labeled samples. We empirically demonstrate that pseudo-labeling can in fact be competitive with the state-of-the-art, while being more resilient to out-of-distribution samples in the unlabeled set. We identify two key factors that allow pseudo-labeling to achieve such remarkable results (1) applying curriculum learning principles and (2) avoiding concept drift by restarting model parameters before each self-training cycle. We obtain 94.91% accuracy on CIFAR-10 using only 4,000 labeled samples, and 68.87% top-1 accuracy on Imagenet-ILSVRC using only 10% of the labeled samples. The code is available at following https URL</p>

<h4 id="code"><a href="https://github.com/uvavision/Curriculum-Labeling">code</a></h4>

<h3 id="citations">Citations</h3>

<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre class="highlight"><code>@misc{grigsby2020measuring,
      title={Measuring Visual Generalization in Continuous Control from Pixels}, 
      author={Jake Grigsby and Yanjun Qi},
      year={2020},
      eprint={2010.06740},
      archivePrefix={arXiv},
      primaryClass={cs.LG}
}
</code></pre></div></div>

<h3 id="support-or-contact">Support or Contact</h3>

<p>Having trouble with our tools? Please <a href="mailto:jcg6dn@virginia.edu">contact Jake</a> and we’ll help you sort it out.</p>]]></content><author><name>Qdata</name></author><category term="AIself" /><category term="Semi-SelfLabel" /><summary type="html"><![CDATA[Title: Curriculum Labeling- Self-paced Pseudo-Labeling for Semi-Supervised Learning”]]></summary></entry><entry><title type="html">NeurIPS - Measuring Visual Generalization in Continuous Control from Pixels</title><link href="https://qdata.github.io//qdata-page/aiself/DMCmaster/" rel="alternate" type="text/html" title="NeurIPS - Measuring Visual Generalization in Continuous Control from Pixels" /><published>2020-12-11T00:00:00-05:00</published><updated>2020-12-11T00:00:00-05:00</updated><id>https://qdata.github.io//qdata-page/aiself/DMCmaster</id><content type="html" xml:base="https://qdata.github.io//qdata-page/aiself/DMCmaster/"><![CDATA[<p><a name="RL-Generalization"></a></p>

<h3 id="title-measuring-visual-generalization-in-continuous-control-from-pixels">Title: Measuring Visual Generalization in Continuous Control from Pixels</h3>

<ul>
  <li>authors: Jake Grigsby, Yanjun Qi</li>
</ul>

<h3 id="paper-arxiv">Paper <a href="https://arxiv.org/abs/2010.06740">Arxiv</a></h3>

<h3 id="code-here">Code <a href="https://github.com/QData/dmc_remastered">Here</a></h3>

<h3 id="abstract">Abstract</h3>
<p>Self-supervised learning and data augmentation have significantly reduced the performance gap between state and image-based reinforcement learning agents in continuous control tasks. However, it is still unclear whether current techniques can face a variety of visual conditions required by real-world environments. We propose a challenging benchmark that tests agents’ visual generalization by adding graphical variety to existing continuous control domains. Our empirical analysis shows that current methods struggle to generalize across a diverse set of visual changes, and we examine the specific factors of variation that make these tasks difficult. We find that data augmentation techniques outperform self-supervised learning approaches and that more significant image transformations provide better visual generalization \footnote{The benchmark and our augmented actor-critic implementation are open-sourced @ this https URL)</p>

<h3 id="citations">Citations</h3>

<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre class="highlight"><code>@misc{grigsby2020measuring,
      title={Measuring Visual Generalization in Continuous Control from Pixels}, 
      author={Jake Grigsby and Yanjun Qi},
      year={2020},
      eprint={2010.06740},
      archivePrefix={arXiv},
      primaryClass={cs.LG}
}
</code></pre></div></div>

<h3 id="support-or-contact">Support or Contact</h3>

<p>Having trouble with our tools? Please <a href="mailto:jcg6dn@virginia.edu">contact Jake</a> and we’ll help you sort it out.</p>]]></content><author><name>Qdata</name></author><category term="AIself" /><category term="RL-Generalization" /><category term="multi-transfer" /><summary type="html"><![CDATA[]]></summary></entry><entry><title type="html">EMNLP - Benchmarking Search Algorithms for Generating NLP Adversarial Examples</title><link href="https://qdata.github.io//qdata-page/aitrust/aiself/benchmark-search/" rel="alternate" type="text/html" title="EMNLP - Benchmarking Search Algorithms for Generating NLP Adversarial Examples" /><published>2020-10-01T00:00:00-04:00</published><updated>2020-10-01T00:00:00-04:00</updated><id>https://qdata.github.io//qdata-page/aitrust/aiself/benchmark-search</id><content type="html" xml:base="https://qdata.github.io//qdata-page/aitrust/aiself/benchmark-search/"><![CDATA[<h3 id="title-searching-for-a-search-method-benchmarking-search-algorithms-for-generating-nlp-adversarial-examples">Title: Searching for a Search Method: Benchmarking Search Algorithms for Generating NLP Adversarial Examples</h3>

<ul>
  <li>
    <p>Abstract:  We study the behavior of several black-box search algorithms used for generating adversarial examples for natural language processing (NLP) tasks. We perform a fine-grained analysis of three elements relevant to search: search algorithm, search space, and search budget. When new search methods are proposed in past work, the attack search space is often modified alongside the search method. Without ablation studies benchmarking the search algorithm change with the search space held constant, an increase in attack success rate could from an improved search method or a less restrictive search space. Additionally, many previous studies fail to properly consider the search algorithms’ run-time cost, which is essential for downstream tasks like adversarial training. Our experiments provide a reproducible benchmark of search algorithms across a variety of search spaces and query budgets to guide future research in adversarial NLP. Based on our experiments, we recommend greedy attacks with word importance ranking when under a time constraint or attacking long inputs, and either beam search or particle swarm optimization otherwise.</p>
  </li>
  <li>
    <p>Citations:</p>
    <div class="language-plaintext highlighter-rouge"><div class="highlight"><pre class="highlight"><code>@misc{yoo2020searching,
    title={Searching for a Search Method: Benchmarking Search Algorithms for Generating NLP Adversarial Examples}, 
    author={Jin Yong Yoo and John X. Morris and Eli Lifland and Yanjun Qi},
    year={2020},
    eprint={2009.06368},
    archivePrefix={arXiv},
    primaryClass={cs.CL}
}
</code></pre></div>    </div>
  </li>
</ul>

<h3 id="our-paper-in-emnlp-blackboxnlp">Our Paper in <a href="https://arxiv.org/abs/2009.06368">EMNLP BlackBoxNLP</a>.</h3>

<h3 id="our-search-benchmarking-result-github---httpsgithubcomqdatatextattack-search-benchmark">Our search benchmarking result Github :  <a href="https://github.com/QData/TextAttack-Search-Benchmark">https://github.com/QData/TextAttack-Search-Benchmark</a></h3>

<h3 id="benchmarking-attack-recipes">Benchmarking Attack Recipes</h3>

<ul>
  <li>
    <p>As we emphasized in the above paper, we don’t recommend to directly compare Attack Recipes out of the box.</p>
  </li>
  <li>
    <p>This is due to that attack recipes in the recent literature used different ways or thresholds in setting up their constraints. Without the constraint space held constant, an increase in attack success rate could come from an improved search or transformation method or a less restrictive search space.</p>
  </li>
</ul>]]></content><author><name>Qdata</name></author><category term="AItrust" /><category term="AIself" /><category term="1-Evasion" /><category term="4-VisualizeBench" /><category term="Generate-Text" /><category term="ToolKit" /><category term="textattack" /><summary type="html"><![CDATA[Title: Searching for a Search Method: Benchmarking Search Algorithms for Generating NLP Adversarial Examples]]></summary></entry></feed>