Home Projects
Projects
Cancel

Projects

Synthesis of animal facial expressions using StarGAN_v2

Synthesis of animal facial expressions using Stargan-v2

LINK

Image Generation
Object Detection

ef

Design


This project focuses on creating personalized pet emojis by detecting dog expressions with EfficientDet and generating corresponding emojis with StarGAN v2. This allows users to quickly transform dog photos into expressive emojis based on detected emotions (anger, happiness, sadness, and yawning).

Data


  • Dataset: Collected approximately 5,000 dog photos using emotion-based keywords [‘anger’, ‘happiness’, ‘sadness’, ‘yawning’].
  • Annotation: Annotated 400 images per emotion for front-facing dog faces, ensuring unique images without duplicates.
  • Image Preprocessing: Cropped images to focus on dog faces, with a minimum resolution of 100x100 pixels for both width and height.

Main model / Technique


Object Detection
EfficientDet was chosen as the object detection model for its accuracy and parameter efficiency, leveraging the EfficientNet backbone.
ef

Image Generation
StarGAN_v2 was used for image-to-image translation across multiple domains, which allowed transformations among four target expressions.
ef

Consideration


ef
Here is the result after image generation using StarGAN v2.

ef
While StarGAN v2 produced varied expressions, a challenge arose: the model also altered the dog’s breed and color based on the reference domain image.

To preserve the breed and color while altering only the facial expression, we integrated Histogram Loss from and color in the generated images. HistoGAN, which ensured consistent breed

ef
After adding Histogram Loss, the results maintained the original breed and color, with only the facial expression changing.

Future Ideas


Considering using an auto-annotation tool, such as CVAT, to improve annotation efficiency and accuracy.

Establishment of Knowledge Graphs using LLM

Establishment of Knowledge Graphs using LLM

LLM
Knowledge Graph (KG)
Named Entity Recognition (NER)
Relation Extraction (RE)

ef

Design


This project aims to automate the extraction of entity relationships within reports on South Korea’s development policies and international cooperation, creating a knowledge graph to improve usability. By automatically identifying entities (such as events, policies, and institutions) and mapping their relationships, the project leverages advanced methodologies to enhance automation efficiency and accuracy in knowledge graph construction.

Data


  • Reports from KDI (Eng)
  • CoNLL 2003 (Eng)

    Preprocessing

  • Coreference Resolution (Stanford CoreNLP)
  • PDF to Text (Tika)

Main model / Technique


Named Entity Recognition (NER)

  • First, We manually BIO tagged all the named entity with TextAE
  • Then we trained We trained BERT+CRF model to detect Named Entity from sentences.

The reason we use CRF(Conditional Random Field) together with BERT
CRF allows to have proper B-I-O sequential structure. It learns the pattern so that we don’t get B-O-I or I-B-O, but it ensures that we always get the right structure of B-I-O. Also I-ORG can’t be followed after B-PER because they are the different entity.

Relation Extraction (RE)
We manually tagged all the relationship between the entities based on Semeval-2010 task. And trained R-BERT which is an enriching Pre-trained Language Model with Entity Information for Relation Classification.

Knowled Graph (KG)
We used Neo4j to build a Knowledge Graph and display the KG.

Consideration


  • The lack of tagged Named Entities and Relation Entities leads to class imbalance in the dataset, resulting in insufficient model training.

Result


ef

Future Ideas


Considering using a better version of NER model such as GliNER

Anomaly Detection for Welding Defect Identification

Anomaly Detection for Welding Defect Identification

Anomaly Detection
Feature Engineering

ef

Design


Performed extensive feature engineering to extract and select key features from sensor data, optimizing inputs for anomaly detection models using clustering techniques, CNN and isolation-based methods, achieving 99% accuracy in identifying welding defects.

Data


  • Time-series seonsor data generated in laser welding production

Main model / Technique


Isolation Forest
Local Outlier Factor (LOF)
Support Vector Machine (SVM)
CNN
STFT / FFT
We applied the Short-time Fourier Transform (STFT) technique to analyze non-stationary signals by segmenting them into narrow time intervals. This transformation allowed us to convert our time series data into an STFT representation, making patterns more discernible. After transforming the data into STFT images, we trained a CNN model to effectively capture and learn the patterns within the time-series data.

Consideration


In AI-based time-series anomaly detection research for automated monitoring, one of the biggest challenges is setting the anomaly score threshold to distinguish between normal and abnormal conditions in real-world applications. This is a crucial task that, when done manually, requires significant expertise and time. Many existing anomaly detection approaches use fixed thresholds, but these can result in high false alarm rates or low anomaly detection rates. To establish an optimal threshold, we focused on clear communication and collaboration, not only with the data analysis team but also with the on-site engineering team, who directly manage the equipment.

Result


Led the end-to-end deployment process, including model development, real-time prediction, and UI integration, successfully implementing a real-time defect detection monitoring system in production environments.

Future Ideas


Time Series Regression for Electrode Surface Temperature Optimization

Time Series Regression for Electrode Surface Temperature Optimization

Regression Modeling
Customized Deep Learning

Design


  • Developed and deployed an ML-based predictive model using time series regression techniques to optimize the drying process in battery manufacturing, leveraging 1D CNN, LSTM, and Attention mechanisms for enhanced accuracy.

Data


  • Time-series data generated in production

Main model / Technique


LSTM
1D-CNN
Attention

Consideration & Thoughts


This project consisted of two main parts: modeling and deployment. Unlike other projects, this one was particularly challenging because we had to define the dataset structure specifically for training the model. The data we received from the equipment was in millisecond intervals, but there were key control points and timing requirements for the model deployment that needed to be considered.

It wasn’t just about building a model with a fixed dataset; we had to determine how to structure the dataset to highlight which points would be most informative and impactful for our model. To explore various dataset configurations, I experimented by stacking multiple neural networks to find the optimal model structure. For example, instead of feeding the entire dataset into a single network input, I separated features based on their importance. When I wanted the model to focus on a specific feature, I created two input streams, feeding less critical features into an initial LSTM layer and later concatenating the more important feature data with the output of this layer.

This project allowed me to experiment extensively with different modeling and data engineering techniques, making it both challenging and enjoyable.

Result


  • Successfully implemented the solution at the Ultium Cells plant (GM-LG joint venture) in Ohio, USA, achieving a significant increase in target temperature accuracy from 50% to 95%.
  • Conducted ongoing remote monitoring and troubleshooting to fine-tune the model, leading to an estimated annual cost savings of $100,000.