<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Projects | Tianma's Website</title><link>https://tianma.netlify.app/project/</link><atom:link href="https://tianma.netlify.app/project/index.xml" rel="self" type="application/rss+xml"/><description>Projects</description><generator>Wowchemy (https://wowchemy.com)</generator><language>en-us</language><lastBuildDate>Thu, 17 Aug 2023 00:00:00 +0000</lastBuildDate><image><url>https://tianma.netlify.app/media/icon_hu0b7a4cb9992c9ac0e91bd28ffd38dd00_9727_512x512_fill_lanczos_center_3.png</url><title>Projects</title><link>https://tianma.netlify.app/project/</link></image><item><title>AI Mahjong</title><link>https://tianma.netlify.app/project/mahjong/</link><pubDate>Thu, 17 Aug 2023 00:00:00 +0000</pubDate><guid>https://tianma.netlify.app/project/mahjong/</guid><description/></item><item><title>Fish2Mesh</title><link>https://tianma.netlify.app/project/fish2mesh/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://tianma.netlify.app/project/fish2mesh/</guid><description/></item><item><title>Home Design</title><link>https://tianma.netlify.app/project/floorplan/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://tianma.netlify.app/project/floorplan/</guid><description>&lt;p>This project is processing during I worked at B&amp;amp;Q company.&lt;/p>
&lt;p>At B&amp;amp;Q, where I previously worked as an Applied Scientist, I built a recommendation system that generated architectural layout suggestions based on user preferences,
transforming floor plans into 2D matrices and enhancing customer experiences through similarity detection algorithms.
This experience reinforced my capabilities in ML model development and demonstrated my capacity for impactful AI-driven solutions.&lt;/p>
&lt;h3 id="step-1-floorplan-detection">Step 1: Floorplan Detection&lt;/h3>
&lt;p>The initial phase focuses on detecting key elements in the floorplan, including spaces, scale marks, walls, doors, and windows.&lt;/p>
&lt;p>Space and Scale Mark Detection: Achieved using a YOLO-based detection model.
Wall, Door, and Window Detection: Handled through a regression network.
Examples of floorplan detection results:
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img src="./example2.png" alt="FloorPlan Detection Result" loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img src="./example1.jpg" alt="FloorPlan Detection Result" loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;h3 id="step-2-space-matching-and-case-recommendation">Step 2: Space Matching and Case Recommendation&lt;/h3>
&lt;p>For each detected space, we match the top 5 design cases tailored to the consumer&amp;rsquo;s preferences.&lt;/p>
&lt;p>Filtering Options: Users can filter recommendations by style and color labels, which are included in our dataset.
Space Matching: To determine the best match, we propose a heatmap of polygon patterns for the floorplan.
This allows us to compute similarity between each space in the floorplan and spaces in our dataset.
Examples of space matching visualizations:
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img src="./cat.png" alt="FloorPlan Detection Result" loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img src="./heatmap.png" alt="FloorPlan Detection Result" loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p></description></item><item><title>Image Compression for classification, object detection and segmentation</title><link>https://tianma.netlify.app/project/icm/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://tianma.netlify.app/project/icm/</guid><description>&lt;p>Recent years have witnessed great advances in deep learning-based image compression, also known as learned image compression. An accurate entropy model is essential in learned image compression, since it can compress high-quality images with a lower bit rate.&lt;/p>
&lt;p>Current learned image compression schemes developed entropy models using context models and hyperpriors. Context models utilize local correlations within latent representations for better probability distribution approximation, while hyperpriors provide side information to estimate distribution parameters. Most recently, several transformer-based learned image compression algorithms have emerged and achieved state-of-the-art rate distortion performances, surpassing existing convolutional neural network (CNN)-based learned image compression and traditional image compression. Transformers are better at modeling long-distance dependencies and extracting global features than CNNs.&lt;/p>
&lt;p>However, the research of transformer-based image compression is still in its early stage. In this work, we propose a novel transformer-based learned image compression model. It adopts transformer structures in the main image encoder and decoder and in the context model. In
particular, we propose a transformer-based spatial-channel auto-regressive context model. Encoded latent-space features are split into spatial-channel chunks, which are entropy encoded sequentially in a channel-first order, followed by a 2D zigzag spatial order, conditioned on previously decoded feature chunks. To reduce the computational complexity, we also adopt a sliding window to restrict the number of chunks participating in the entropy model. Experimental studies on public image compression datasets demonstrate that our proposed transformer-based learned image codec outperforms traditional image compression and existing learned image compression models visually and quantitatively.&lt;/p>
&lt;p>The more details are show up in the Github link: &lt;a href="https://github.com/stm233/image-compression-with-swin-transformer" target="_blank" rel="noopener">https://github.com/stm233/image-compression-with-swin-transformer&lt;/a>.&lt;/p></description></item><item><title>Image Compression with Swin Transformer</title><link>https://tianma.netlify.app/project/image-compression/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://tianma.netlify.app/project/image-compression/</guid><description>&lt;p>Recent years have witnessed great advances in deep learning-based image compression, also known as learned image compression. An accurate entropy model is essential in learned image compression, since it can compress high-quality images with a lower bit rate.&lt;/p>
&lt;p>Current learned image compression schemes developed entropy models using context models and hyperpriors. Context models utilize local correlations within latent representations for better probability distribution approximation, while hyperpriors provide side information to estimate distribution parameters. Most recently, several transformer-based learned image compression algorithms have emerged and achieved state-of-the-art rate distortion performances, surpassing existing convolutional neural network (CNN)-based learned image compression and traditional image compression. Transformers are better at modeling long-distance dependencies and extracting global features than CNNs.&lt;/p>
&lt;p>However, the research of transformer-based image compression is still in its early stage. In this work, we propose a novel transformer-based learned image compression model. It adopts transformer structures in the main image encoder and decoder and in the context model. In
particular, we propose a transformer-based spatial-channel auto-regressive context model. Encoded latent-space features are split into spatial-channel chunks, which are entropy encoded sequentially in a channel-first order, followed by a 2D zigzag spatial order, conditioned on previously decoded feature chunks. To reduce the computational complexity, we also adopt a sliding window to restrict the number of chunks participating in the entropy model. Experimental studies on public image compression datasets demonstrate that our proposed transformer-based learned image codec outperforms traditional image compression and existing learned image compression models visually and quantitatively.&lt;/p>
&lt;p>The more details are show up in the Github link: &lt;a href="https://github.com/stm233/image-compression-with-swin-transformer" target="_blank" rel="noopener">https://github.com/stm233/image-compression-with-swin-transformer&lt;/a>.&lt;/p></description></item><item><title>MAGAN</title><link>https://tianma.netlify.app/project/magan/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://tianma.netlify.app/project/magan/</guid><description>&lt;p>Generative Adversarial Networks (GANs) suffer from the instability of training because of Optimal Transportation (OT) problems.
Based on Brenier&amp;rsquo;s Theorem, we converted OT problems into solving the elliptic Monge–Ampère Partial Differential Equation (MAPDE) by utilizing the finite difference method.
In order to solve n (n &amp;gt; 3) dimensional MAPDE, we improved Neumann boundary conditions and extended a discretization of MAPDE for the numerical solution to enable the optimal map between generators and discriminators.
The solution of MAPDE was regarded as a new divergence instead of Wasserstein Distance from WGAN.&lt;/p>
&lt;p>We provided several computational examples to demonstrate that the precision was increased by 5.3%.
Moreover, MAGAN was able to stabilize training with almost no hyperparameter tuning and the convergent speed of MAGAN was 317.2% faster than WGAN-GP on LSUN Bedrooms Database.
MAGAN also achieved the Inception Score (IS) of 8.7 on CIFAR-10.&lt;/p></description></item></channel></rss>