Sitemap

A list of all the posts and pages found on the site. For you robots out there is an XML version available for digesting as well.

Pages

Posts

Future Blog Post

less than 1 minute read

Published:

This post will show up by default. To disable scheduling of future posts, edit config.yml and set future: false.

Blog Post number 4

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 3

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 2

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 1

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

portfolio

publications

Pretrained Language Models are Symbolic Mathematics Solvers too!

Published in arxiv, 2021

Solving symbolic mathematics has always been of in the arena of human ingenuity that needs compositional reasoning and recurrence. However, recent studies have shown that large-scale language models such as transformers are universal and surprisingly can be trained as a sequence-to-sequence task to solve complex mathematical equations. These large transformer models need humongous amounts of training data to generalize to unseen symbolic mathematics problems. In this paper, we present a sample efficient way of solving the symbolic tasks by first pretraining the transformer model with language translation and then fine-tuning the pretrained transformer model to solve the downstream task of symbolic mathematics. We achieve comparable accuracy on the integration task with our pretrained model while using around 1.5 orders of magnitude less number of training samples with respect to the state-of-the-art deep learning for symbolic mathematics. The test accuracy on differential equation tasks is considerably lower comparing with integration as they need higher order recursions that are not present in language translations. We pretrain our model with different pairs of language translations. Our results show language bias in solving symbolic mathematics tasks. Finally, we study the robustness of the fine-tuned model on symbolic math tasks against distribution shift, and our approach generalizes better in distribution shift scenarios for the function integration.

Paper.Code

Recommended citation: Noorbakhsh, Kimia, Modar Sulaiman, Mahdi Sharifi, Kallol Roy, and Pooyan Jamshidi. "Pretrained Language Models are Symbolic Mathematics Solvers too!." arXiv preprint arXiv:2110.03501 (2021).

LPF-Defense: 3D Adversarial Defense based on Frequency Analysis

Published in PLOS ONE, 2022

Although 3D point cloud classification has recently been widely deployed in different application scenarios, it is still very vulnerable to adversarial attacks. This increases the importance of robust training of 3D models in the face of adversarial attacks. Based on our analysis on the performance of existing adversarial attacks, more adversarial perturbations are found in the mid and high-frequency components of input data. Therefore, by suppressing the high-frequency content in the training phase, the models robustness against adversarial examples is improved. Experiments showed that the proposed defense method decreases the success rate of six attacks on PointNet, PointNet++, and DGCNN models. In particular, improvements are achieved with an average increase of classification accuracy by 3.8 % on drop100 attack and 4.26 % on drop200 attack compared to the state-of-the-art methods. The method also improves models accuracy on the original dataset compared to other available methods.

Paper.Code

Recommended citation: Naderi, H., Noorbakhsh, K., Etemadi, A., & Kasaei, S. (02 2023). LPF-Defense: 3D adversarial defense based on frequency analysis. PLOS ONE, 18(2), 1–19. doi:10.1371/journal.pone.0271388

Counterfactual Temporal Point Processes

Published in Advances in Neural Information Processing Systems (NeurIPS), 2022

Machine learning models based on temporal point processes are the state of the art in a wide variety of applications involving discrete events in continuous time. However, these models lack the ability to answer counterfactual questions, which are increasingly relevant as these models are being used to inform targeted interventions. In this work, our goal is to fill this gap. To this end, we first develop a causal model of thinning for temporal point processes that builds upon the Gumbel-Max structural causal model. This model satisfies a desirable counterfactual monotonicity condition, which is sufficient to identify counterfactual dynamics in the process of thinning. Then, given an observed realization of a temporal point process with a given intensity function, we develop a sampling algorithm that uses the above causal model of thinning and the superposition theorem to simulate counterfactual realizations of the temporal point process under a given alternative intensity function. Simulation experiments using synthetic and real epidemiological data show that the counterfactual realizations provided by our algorithm may give valuable insights to enhance targeted interventions.

Paper.Code.Talk

Recommended citation: Noorbakhsh, Kimia, and Manuel Gomez Rodriguez. "Counterfactual Temporal Point Processes." In Advances in Neural Information Processing Systems (2022).

Published in , 1900

TACO: Vision Models Can Be Efficiently Specialized via Few-Shot Task-Aware Compression

Published in arxiv, 2023

Recent vision architectures and self-supervised training methods enable vision models that are extremely accurate and general, but come with massive parameter and computational costs. In practical settings, such as camera traps, users have limited resources, and may fine-tune a pretrained model on (often limited) data from a small set of specific categories of interest. These users may wish to make use of modern, highly-accurate models, but are often computationally constrained. To address this, we ask: can we quickly compress large generalist models into accurate and efficient specialists? For this, we propose a simple and versatile technique called Few-Shot Task-Aware Compression (TACO). Given a large vision model that is pretrained to be accurate on a broad task, such as classification over ImageNet-22K, TACO produces a smaller model that is accurate on specialized tasks, such as classification across vehicle types or animal species. Crucially, TACO works in few-shot fashion, i.e. only a few task-specific samples are used, and the procedure has low computational overheads. We validate TACO on highly-accurate ResNet, ViT/DeiT, and ConvNeXt models, originally trained on ImageNet, LAION, or iNaturalist, which we specialize and compress to a diverse set of “downstream” subtasks. TACO can reduce the number of non-zero parameters in existing models by up to 20x relative to the original models, leading to inference speedups of up to 3×, while remaining accuracy-competitive with the uncompressed models on the specialized tasks.

Paper**

Recommended citation: D Kuznedelev*, S Tabesh*, K Noorbakhsh*, E Frantar*, S Beery, E Kurtic, D Alistarh. "Vision Models Can Be Efficiently Specialized via Few-Shot Task-Aware Compression." arXiv preprint arXiv:2303.14409 (2023).

talks

teaching

Scientific and Technical Presentation

CE-40221, Sharif University, CE Department (Fall 2021, Spring 2022)

Designed and graded course assignments regarding Business Letter writing, Resume writing and Scientific Paper writing. Designed and graded midterm and final exam questions.

Design of Algorithms

CE-40354, Sharif University, CE Department (Fall 2021 (Head TA), Spring 2021 (Head of Assignments))

Led and managed a group of grad and undergrad TA’s to design and grade course assignments, design and grade course exams, and hold TA sessions during the semester. [Fall Course website]

Discrete Structures

CE-40115, Sharif University, CE Department (Spring 2022 (Head TA), Spring 2021 (Head TA), Spring 2020)

Led and managed a group of grad and undergrad TA’s to design and grade course assignments, design and grade course exams, and hold TA sessions during the semester. [2022 Course website]

Artificial Intelligence

CE-40417, Sharif University, CE Department (Spring 2022, Fall 2021)

Designed and graded Machine Learning and Deep Learning Assignments (theoretical and practical). Also designed and graded the course’s final exam questions regarding ML and DL. [Course website]