Skip to main content

Showing 1–13 of 13 results for author: Smirnov, D

Searching in archive cs. Search in all archives.
.
  1. arXiv:2511.17844  [pdf, ps, other

    cs.CV cs.AI

    Less is More: Data-Efficient Adaptation for Controllable Text-to-Video Generation

    Authors: Shihan Cheng, Nilesh Kulkarni, David Hyde, Dmitriy Smirnov

    Abstract: Fine-tuning large-scale text-to-video diffusion models to add new generative controls, such as those over physical camera parameters (e.g., shutter speed or aperture), typically requires vast, high-fidelity datasets that are difficult to acquire. In this work, we propose a data-efficient fine-tuning strategy that learns these controls from sparse, low-quality synthetic data. We show that not only… ▽ More

    Submitted 10 December, 2025; v1 submitted 21 November, 2025; originally announced November 2025.

    MSC Class: 68U05 ACM Class: I.3.3; I.5.4

  2. arXiv:2506.09440  [pdf, ps, other

    cs.CL cs.AI

    GigaChat Family: Efficient Russian Language Modeling Through Mixture of Experts Architecture

    Authors: GigaChat team, Mamedov Valentin, Evgenii Kosarev, Gregory Leleytner, Ilya Shchuckin, Valeriy Berezovskiy, Daniil Smirnov, Dmitry Kozlov, Sergei Averkiev, Lukyanenko Ivan, Aleksandr Proshunin, Ainur Israfilova, Ivan Baskov, Artem Chervyakov, Emil Shakirov, Mikhail Kolesov, Daria Khomich, Darya Latortseva, Sergei Porkhun, Yury Fedorov, Oleg Kutuzov, Polina Kudriavtseva, Sofiia Soldatova, Kolodin Egor, Stanislav Pyatkin , et al. (9 additional authors not shown)

    Abstract: Generative large language models (LLMs) have become crucial for modern NLP research and applications across various languages. However, the development of foundational models specifically tailored to the Russian language has been limited, primarily due to the significant computational resources required. This paper introduces the GigaChat family of Russian LLMs, available in various sizes, includi… ▽ More

    Submitted 11 June, 2025; originally announced June 2025.

    Comments: ACL-2025 System Demo

  3. arXiv:2411.01212  [pdf, other

    cs.CV cs.AI cs.GR cs.LG

    Infinite-Resolution Integral Noise Warping for Diffusion Models

    Authors: Yitong Deng, Winnie Lin, Lingxiao Li, Dmitriy Smirnov, Ryan Burgert, Ning Yu, Vincent Dedun, Mohammad H. Taghavi

    Abstract: Adapting pretrained image-based diffusion models to generate temporally consistent videos has become an impactful generative modeling research direction. Training-free noise-space manipulation has proven to be an effective technique, where the challenge is to preserve the Gaussian white noise distribution while adding in temporal consistency. Recently, Chang et al. (2024) formulated this problem u… ▽ More

    Submitted 2 November, 2024; originally announced November 2024.

  4. arXiv:2306.13702  [pdf, other

    cs.GR

    Magenta Green Screen: Spectrally Multiplexed Alpha Matting with Deep Colorization

    Authors: Dmitriy Smirnov, Chloe LeGendre, Xueming Yu, Paul Debevec

    Abstract: We introduce Magenta Green Screen, a novel machine learning--enabled matting technique for recording the color image of a foreground actor and a simultaneous high-quality alpha channel without requiring a special camera or manual keying techniques. We record the actor on a green background but light them with only red and blue foreground lighting. In this configuration, the green channel shows the… ▽ More

    Submitted 23 June, 2023; originally announced June 2023.

    Comments: In DigiPro 2023

  5. arXiv:2201.11940  [pdf, other

    cs.GR cs.AI

    Wassersplines for Neural Vector Field--Controlled Animation

    Authors: Paul Zhang, Dmitriy Smirnov, Justin Solomon

    Abstract: Much of computer-generated animation is created by manipulating meshes with rigs. While this approach works well for animating articulated objects like animals, it has limited flexibility for animating less structured free-form objects. We introduce Wassersplines, a novel trajectory inference method for animating unstructured densities based on recent advances in continuous normalizing flows and o… ▽ More

    Submitted 19 September, 2022; v1 submitted 28 January, 2022; originally announced January 2022.

  6. arXiv:2111.09383  [pdf, other

    cs.CV cs.GR

    DeepCurrents: Learning Implicit Representations of Shapes with Boundaries

    Authors: David Palmer, Dmitriy Smirnov, Stephanie Wang, Albert Chern, Justin Solomon

    Abstract: Recent techniques have been successful in reconstructing surfaces as level sets of learned functions (such as signed distance fields) parameterized by deep neural networks. Many of these methods, however, learn only closed surfaces and are unable to reconstruct shapes with boundary curves. We propose a hybrid shape representation that combines explicit boundary curves with implicit learned interio… ▽ More

    Submitted 21 March, 2022; v1 submitted 17 November, 2021; originally announced November 2021.

  7. Interactive All-Hex Meshing via Cuboid Decomposition

    Authors: Lingxiao Li, Paul Zhang, Dmitriy Smirnov, S. Mazdak Abulnaga, Justin Solomon

    Abstract: Standard PolyCube-based hexahedral (hex) meshing methods aim to deform the input domain into an axis-aligned PolyCube volume with integer corners; if this deformation is bijective, then applying the inverse map to the voxelized PolyCube yields a valid hex mesh. A key challenge in these methods is to maintain the bijectivity of the PolyCube deformation, thus reducing the robustness of these algorit… ▽ More

    Submitted 13 September, 2021; originally announced September 2021.

    ACM Class: I.3.5

  8. arXiv:2104.14553  [pdf, other

    cs.CV

    MarioNette: Self-Supervised Sprite Learning

    Authors: Dmitriy Smirnov, Michael Gharbi, Matthew Fisher, Vitor Guizilini, Alexei A. Efros, Justin Solomon

    Abstract: Artists and video game designers often construct 2D animations using libraries of sprites -- textured patches of objects and characters. We propose a deep learning approach that decomposes sprite-based video animations into a disentangled representation of recurring graphic elements in a self-supervised manner. By jointly learning a dictionary of possibly transparent patches and training a network… ▽ More

    Submitted 20 October, 2021; v1 submitted 29 April, 2021; originally announced April 2021.

    Comments: Accepted to NeurIPS 2021

  9. arXiv:2104.12826  [pdf, other

    cs.GR

    HodgeNet: Learning Spectral Geometry on Triangle Meshes

    Authors: Dmitriy Smirnov, Justin Solomon

    Abstract: Constrained by the limitations of learning toolkits engineered for other applications, such as those in image processing, many mesh-based learning algorithms employ data flows that would be atypical from the perspective of conventional geometry processing. As an alternative, we present a technique for learning from meshes built from standard geometry processing modules and operations. We show that… ▽ More

    Submitted 26 April, 2021; originally announced April 2021.

    Comments: Accepted to SIGGRAPH 2021

  10. arXiv:2004.14875  [pdf, other

    cs.CV cs.LG eess.IV

    Polygonal Building Segmentation by Frame Field Learning

    Authors: Nicolas Girard, Dmitriy Smirnov, Justin Solomon, Yuliya Tarabalka

    Abstract: While state of the art image segmentation models typically output segmentations in raster format, applications in geographic information systems often require vector polygons. To help bridge the gap between deep network output and the format used in downstream tasks, we add a frame field output to a deep segmentation model for extracting buildings from remote sensing images. We train a deep neural… ▽ More

    Submitted 31 March, 2021; v1 submitted 30 April, 2020; originally announced April 2020.

    Comments: CVPR 2021 - IEEE Conference on Computer Vision and Pattern Recognition, Jun 2021, Pittsburg / Virtual, United States

    Report number: hal-02548545, v2

  11. arXiv:1906.12337  [pdf, other

    cs.GR

    Learning Manifold Patch-Based Representations of Man-Made Shapes

    Authors: Dmitriy Smirnov, Mikhail Bessmeltsev, Justin Solomon

    Abstract: Choosing the right representation for geometry is crucial for making 3D models compatible with existing applications. Focusing on piecewise-smooth man-made shapes, we propose a new representation that is usable in conventional CAD modeling pipelines and can also be learned by deep neural networks. We demonstrate its benefits by applying it to the task of sketch-based modeling. Given a raster image… ▽ More

    Submitted 9 February, 2021; v1 submitted 28 June, 2019; originally announced June 2019.

    Comments: Accepted to ICLR 2021

  12. arXiv:1904.08921  [pdf, other

    cs.GR cs.CV

    Deep Parametric Shape Predictions using Distance Fields

    Authors: Dmitriy Smirnov, Matthew Fisher, Vladimir G. Kim, Richard Zhang, Justin Solomon

    Abstract: Many tasks in graphics and vision demand machinery for converting shapes into consistent representations with sparse sets of parameters; these representations facilitate rendering, editing, and storage. When the source data is noisy or ambiguous, however, artists and engineers often manually construct such representations, a tedious and potentially time-consuming process. While advances in deep le… ▽ More

    Submitted 19 March, 2020; v1 submitted 18 April, 2019; originally announced April 2019.

    Comments: Accepted to CVPR 2020

    Journal ref: Proc. of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. (2020) 561-570

  13. arXiv:1701.02595  [pdf

    math.OC cs.SI

    Around power law for PageRank components in Buckley-Osthus model of web graph

    Authors: Alexander Gasnikov, Maxim Zhukovskii, Sergey Kim, Fedor Noskov, Stepan Plaunov, Daniil Smirnov

    Abstract: In the paper we investigate power law for PageRank components for the Buckley-Osthus model for web graph. We compare different numerical methods for PageRank calculation. With the best method we do a lot of numerical experiments. These experiments confirm the hypothesis about power law. At the end we discuss real model of web-ranking based on the classical PageRank approach.

    Submitted 1 March, 2017; v1 submitted 8 January, 2017; originally announced January 2017.

    Comments: in Russian, 41 pages