논문 인용하기
각 논문마다 생성되어 있는 BibTeX를 사용하시면 자신이 원하는 스타일의 인용 문구를 생성할 수 있습니다.
생성된 BibTeX 코드를 복사하여 BibTeX Parser를 사용해 일반 문자열로 바꾸십시오. 아래의 사이트와 같이 웹에서 변환할 수도 있습니다.
bibtex.online2024
안성배,; 이재현,; 박보현,; 오상윤,
HPC 환경에서의 스케줄링을 위한 강화학습 및 휴리스틱 알고리즘의 비교 분석🇰🇷 DomesticConference
2024년도 한국통신학회 동계종합학술발표회, 한국통신학회, 2024.
@conference{kics2024-1,
title = {HPC 환경에서의 스케줄링을 위한 강화학습 및 휴리스틱 알고리즘의 비교 분석},
author = {안성배 and 이재현 and 박보현 and 오상윤},
url = {https://www.dbpia.co.kr/journal/articleDetail?nodeId=NODE11737204},
year = {2024},
date = {2024-03-27},
urldate = {2024-03-27},
booktitle = {2024년도 한국통신학회 동계종합학술발표회},
publisher = {한국통신학회},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Paulo, C. Sergio; 유미리,; 최지헌,; 오상윤,
Dynamic Programming-Based Multilevel Graph Partitioning for Large-Scale Graph Data🇰🇷 DomesticConference
2024년도 한국통신학회 동계종합학술발표회, 한국통신학회, 2024.
Abstract | Links | BibTeX | 태그:
@conference{2024kics-2,
title = {Dynamic Programming-Based Multilevel Graph Partitioning for Large-Scale Graph Data},
author = {C. Sergio Paulo and 유미리 and 최지헌 and 오상윤},
url = {https://www.dbpia.co.kr/journal/articleDetail?nodeId=NODE11737048},
year = {2024},
date = {2024-03-27},
booktitle = {2024년도 한국통신학회 동계종합학술발표회},
publisher = {한국통신학회},
abstract = {Multilevel graph algorithms are used to create optimal partitions for large graphs. However, the dynamic changes to the graph structures during partitioning lead to increased memory. These changes involve adding temporal data to arrays or queues during intermediary operations. To enhance efficiency and minimize memory usage, we integrated dynamic programming. Experimental results demonstrate the improved scalability and effectiveness of the proposed approach in terms of memory usage.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Yoon, Daegun; Oh, Sang-yoon
Preserving Near-Optimal Gradient Sparsification Cost for Scalable Distributed Deep Learning🌏 InternationalConference
The 24th IEEE/ACM international Symposium on Cluster, Cloud and Internet Computing (CCGrid 2024), 2024.
@conference{icpp2024yoon,
title = {Preserving Near-Optimal Gradient Sparsification Cost for Scalable Distributed Deep Learning},
author = {Daegun Yoon and Sang-yoon Oh},
url = {https://arxiv.org/abs/2402.13781},
year = {2024},
date = {2024-02-13},
booktitle = {The 24th IEEE/ACM international Symposium on Cluster, Cloud and Internet Computing (CCGrid 2024)},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
2023
Jung, Hyunseok; Choi, Jiheon; Park, Jongwon; Baek, Sehui; Oh, Sangyoon
Conditional LSTM-VAE-based Data Augmentation for Disaster Classification Prediction🌏 InternationalConference
The 9th International Conference on Next Generation Computing 2023, 2023.
BibTeX | 태그:
@conference{nokey,
title = {Conditional LSTM-VAE-based Data Augmentation for Disaster Classification Prediction},
author = {Hyunseok Jung and Jiheon Choi and Jongwon Park and Sehui Baek and Sangyoon Oh },
year = {2023},
date = {2023-11-24},
urldate = {2023-11-24},
booktitle = {The 9th International Conference on Next Generation Computing 2023},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Yu, Miri; Kwon, Oh-Kyoung; Oh, Sangyoon (Ed.)
Addressing Client Heterogeneity in Synchronous Federated Learning: The CHAFL Approach🌏 InternationalConference
The 29th IEEE International Conference on Parallel and Distributed Systems (ICPADS 2023), 2023.
BibTeX | 태그: federated learning
@conference{nokey,
title = {Addressing Client Heterogeneity in Synchronous Federated Learning: The CHAFL Approach},
editor = {Miri Yu and Oh-Kyoung Kwon and Sangyoon Oh},
year = {2023},
date = {2023-11-10},
urldate = {2023-11-10},
booktitle = {The 29th IEEE International Conference on Parallel and Distributed Systems (ICPADS 2023)},
keywords = {federated learning},
pubstate = {published},
tppubtype = {conference}
}
Yoon, Daegun; Oh, Sangyoon
MiCRO: Near-Zero Cost Gradient Sparsification for Scaling and Accelerating Distributed DNN Training🌏 InternationalConference
30th IEEE International Conference on High Performance Computing, Data, and Analytics (HiPC 2023), 2023.
Links | BibTeX | 태그: distributed deep learning, gradient sparsification
@conference{nokey,
title = {MiCRO: Near-Zero Cost Gradient Sparsification for Scaling and Accelerating Distributed DNN Training},
author = {Daegun Yoon and Sangyoon Oh},
url = {https://arxiv.org/abs/2310.00967},
year = {2023},
date = {2023-10-02},
urldate = {2023-10-02},
booktitle = {30th IEEE International Conference on High Performance Computing, Data, and Analytics (HiPC 2023)},
keywords = {distributed deep learning, gradient sparsification},
pubstate = {published},
tppubtype = {conference}
}
Yoon, Daegun; Oh, Sangyoon
DEFT: Exploiting Gradient Norm Difference between Model Layers for Scalable Gradient Sparsification🌏 InternationalConference
International Conference on Parallel Processing (ICPP) 2023, 2023.
Abstract | Links | BibTeX | 태그: distributed deep learning, gradient sparsification
@conference{nokey,
title = {DEFT: Exploiting Gradient Norm Difference between Model Layers for Scalable Gradient Sparsification},
author = {Daegun Yoon and Sangyoon Oh},
url = {https://arxiv.org/abs/2307.03500},
year = {2023},
date = {2023-08-07},
urldate = {2023-08-07},
booktitle = {International Conference on Parallel Processing (ICPP) 2023},
abstract = {Gradient sparsification is a widely adopted solution for reducing
the excessive communication traffic in distributed deep learning.
However, most existing gradient sparsifiers have relatively poor
scalability because of considerable computational cost of gradient
selection and/or increased communication traffic owing to gradient
build-up. To address these challenges, we propose a novel gradient
sparsification scheme, DEFT, that partitions the gradient selection
task into sub tasks and distributes them to workers. DEFT differs
from existing sparsifiers, wherein every worker selects gradients
among all gradients. Consequently, the computational cost can
be reduced as the number of workers increases. Moreover, gradient build-up can be eliminated because DEFT allows workers to
select gradients in partitions that are non-intersecting (between
workers). Therefore, even if the number of workers increases, the
communication traffic can be maintained as per user requirement.
To avoid the loss of significance of gradient selection, DEFT
selects more gradients in the layers that have a larger gradient
norm than the other layers. Because every layer has a different
computational load, DEFT allocates layers to workers using a binpacking algorithm to maintain a balanced load of gradient selection
between workers. In our empirical evaluation, DEFT shows a significant improvement in training performance in terms of speed
in gradient selection over existing sparsifiers while achieving high
convergence performance.},
keywords = {distributed deep learning, gradient sparsification},
pubstate = {published},
tppubtype = {conference}
}
the excessive communication traffic in distributed deep learning.
However, most existing gradient sparsifiers have relatively poor
scalability because of considerable computational cost of gradient
selection and/or increased communication traffic owing to gradient
build-up. To address these challenges, we propose a novel gradient
sparsification scheme, DEFT, that partitions the gradient selection
task into sub tasks and distributes them to workers. DEFT differs
from existing sparsifiers, wherein every worker selects gradients
among all gradients. Consequently, the computational cost can
be reduced as the number of workers increases. Moreover, gradient build-up can be eliminated because DEFT allows workers to
select gradients in partitions that are non-intersecting (between
workers). Therefore, even if the number of workers increases, the
communication traffic can be maintained as per user requirement.
To avoid the loss of significance of gradient selection, DEFT
selects more gradients in the layers that have a larger gradient
norm than the other layers. Because every layer has a different
computational load, DEFT allocates layers to workers using a binpacking algorithm to maintain a balanced load of gradient selection
between workers. In our empirical evaluation, DEFT shows a significant improvement in training performance in terms of speed
in gradient selection over existing sparsifiers while achieving high
convergence performance.
유미리,; 윤대건,; 오상윤,
연합학습 기법들의 성능평가를 지원하는 이기종 기반의 실험 플랫폼 설계🇰🇷 DomesticConference
2023년도 한국통신학회 하계종합학술발표회 , 한국통신학회 2023.
Links | BibTeX | 태그: federated learning
@conference{연합학습기법들의성능평가를지원하는이기종기반의실험플랫폼설계,
title = {연합학습 기법들의 성능평가를 지원하는 이기종 기반의 실험 플랫폼 설계},
author = {유미리 and 윤대건 and 오상윤},
url = {https://www.dbpia.co.kr/journal/articleDetail?nodeId=NODE11487802},
year = {2023},
date = {2023-06-21},
urldate = {2023-06-21},
booktitle = {2023년도 한국통신학회 하계종합학술발표회 },
organization = { 한국통신학회},
keywords = {federated learning},
pubstate = {published},
tppubtype = {conference}
}
이재현,; 정현석,; 오상윤,
VM 배치를 위한 DDQN 기반 태스크 스케줄링 알고리즘🇰🇷 DomesticConference
2023년도 한국통신학회 하계종합학술발표회 , 한국통신학회 2023.
Links | BibTeX | 태그: deep reinforcement learning
@conference{nokey,
title = {VM 배치를 위한 DDQN 기반 태스크 스케줄링 알고리즘},
author = {이재현 and 정현석 and 오상윤 },
url = {https://www.dbpia.co.kr/journal/articleDetail?nodeId=NODE11487081},
year = {2023},
date = {2023-06-21},
urldate = {2023-06-21},
booktitle = {2023년도 한국통신학회 하계종합학술발표회 },
organization = {한국통신학회},
keywords = {deep reinforcement learning},
pubstate = {published},
tppubtype = {conference}
}
Baek, Minseok; Paulo, C. Sergio; Oh, Sangyoon
Analysis of the In-Memory Checkpointing Approach in Apache Flink🇰🇷 DomesticConference
2023년도 한국통신학회 하계종합학술발표회 , 한국통신학회 2023.
@conference{nokey,
title = {Analysis of the In-Memory Checkpointing Approach in Apache Flink},
author = {Minseok Baek and C. Sergio Paulo and Sangyoon Oh},
url = {https://www.dbpia.co.kr/pdf/pdfView.do?nodeId=NODE11487634},
year = {2023},
date = {2023-06-21},
urldate = {2023-06-21},
booktitle = {2023년도 한국통신학회 하계종합학술발표회 },
organization = { 한국통신학회},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
정현석,; 최지헌,; 박종원,; 백세희,; 오상윤,
화재 감지 시스템을 위한 MLOps 시스템 구조🇰🇷 DomesticConference
2023 한국차세대컴퓨팅학회 춘계학술대회 , 한국차세대컴퓨팅학회 2023.
Links | BibTeX | 태그: edge computing, MLOps
@conference{MLOps,
title = {화재 감지 시스템을 위한 MLOps 시스템 구조},
author = {정현석 and 최지헌 and 박종원 and 백세희 and 오상윤 },
url = {https://www.earticle.net/Article/A433574},
year = {2023},
date = {2023-05-31},
urldate = {2023-05-31},
booktitle = {2023 한국차세대컴퓨팅학회 춘계학술대회 },
pages = {313-315},
organization = {한국차세대컴퓨팅학회 },
keywords = {edge computing, MLOps},
pubstate = {published},
tppubtype = {conference}
}
Lee, Seungjun; Yu, Miri; Yoon, Daegun; Oh, Sangyoon
Can hierarchical client clustering mitigate the data heterogeneity effect in federated learning?🌏 InternationalConference
2023 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW), 2023, ISBN: 979-8-3503-1200-3.
Abstract | Links | BibTeX | 태그: federated learning
@conference{nokey,
title = {Can hierarchical client clustering mitigate the data heterogeneity effect in federated learning?},
author = {Seungjun Lee and Miri Yu and Daegun Yoon and Sangyoon Oh},
url = {10.1109/IPDPSW59300.2023.00134},
doi = {10.1109/IPDPSW59300.2023.00134},
isbn = {979-8-3503-1200-3},
year = {2023},
date = {2023-05-15},
urldate = {2023-05-15},
booktitle = {2023 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW)},
abstract = {Federated learning (FL) was proposed for training a deep neural network model using millions of user data. The technique has attracted considerable attention owing to its privacy-preserving characteristic. However, two major challenges exist. The first is the limitation of simultaneously participating clients. If the number of clients increases, the single parameter server easily becomes a bottleneck and is prone to have stragglers. The second is data heterogeneity, which adversely affects the accuracy of the global model. Because data should remain at user devices to preserve privacy, we cannot use data shuffling, which is used to homogenize training data in traditional distributed deep learning. We propose a client clustering and model aggregation method, CCFed, to increase the number of simultaneously participating clients and mitigate the data heterogeneity problem. CCFed improves the learning performance using set partition modeling to let data be evenly distributed between clusters and mitigate the effect of a non-IID environment. Experiments show that we can achieve a 2.7-14% higher accuracy using CCFed compared with FedAvg, where CCFed requires approximately 50% less number of rounds compared with FedAvg training on benchmark datasets.},
keywords = {federated learning},
pubstate = {published},
tppubtype = {conference}
}
최지헌,; 유미리,; 윤대건,; 오상윤,
연합학습에서의 보안 취약점 분석🇰🇷 DomesticConference
2023년도 한국통신학회 동계종합학술발표회 논문집 , vol. 80, 한국통신학회 2023, ISSN: 2383-8302.
Abstract | Links | BibTeX | 태그: federated learning
@conference{최지헌2023연합학습에서의,
title = {연합학습에서의 보안 취약점 분석},
author = {최지헌 and 유미리 and 윤대건 and 오상윤},
url = {https://www.dbpia.co.kr/pdf/pdfView.do?nodeId=NODE11227811},
issn = {2383-8302},
year = {2023},
date = {2023-02-28},
urldate = {2023-02-28},
booktitle = {2023년도 한국통신학회 동계종합학술발표회 논문집
},
volume = {80},
pages = {1201-1202},
organization = {한국통신학회},
abstract = {개인 데이터에 대한 프라이버시 침해 없이 분산 기계학습을 구현하기 위해 연합학습이 제안되었다. 기존 연합학습 기법의 개선을 통해 정확도향상 및 수렴속도 향상을 목표로 하는 새로운 기법들이 등장하고 있어서, 이에 대한 보안 가이드라인이 필요한 상황이다. 본 논문에서는연합학습 구조의 특징으로 나타나는 보안 취약점을 공격형태 별로 구분하고 이에 대한 대응방안을 고찰한다.},
keywords = {federated learning},
pubstate = {published},
tppubtype = {conference}
}
백민석,; 정현석,; 공은빈,; 오상윤,
자율주행 데이터의 효과적인 처리를 위한 분산 데이터베이스 설계🇰🇷 DomesticConference
2023년도 한국통신학회 동계종합학술발표회 논문집, vol. 80, 한국통신학회 2023, ISBN: 2383-8302.
Abstract | Links | BibTeX | 태그: 분산 데이터베이스, 자율주행
@conference{,
title = {자율주행 데이터의 효과적인 처리를 위한 분산 데이터베이스 설계},
author = {백민석 and 정현석 and 공은빈 and 오상윤},
url = {https://www.dbpia.co.kr/pdf/pdfView.do?nodeId=NODE11227933},
isbn = {2383-8302},
year = {2023},
date = {2023-02-28},
urldate = {2023-02-28},
booktitle = {2023년도 한국통신학회 동계종합학술발표회 논문집},
volume = {80},
pages = {1,411 - 1,412},
organization = {한국통신학회},
abstract = {자율주행 기술 고도화를 위해서는 관련 데이터의 효과적인 관리를 지원하는 시스템이 반드시 필요하다. 본 논문에서는, 비정형 대용량의 자율주행 데이터를 처리하기 위한 HDFS 와 HBase 기반의 분산 데이터베이스의 설계를 소개하며, 공개 자율주행 데이터의 ETL 과정을 통해 실증적인 효과를 분석한다. },
keywords = {분산 데이터베이스, 자율주행},
pubstate = {published},
tppubtype = {conference}
}
Yoon, Daegun; Jeong, Minjoong; Oh, Sangyoon
SAGE: toward on-the-fly gradient compression ratio scaling🌏 InternationalJournal Article
In: The Journal of Supercomputing, pp. 1–23, 2023.
Abstract | Links | BibTeX | 태그: distributed deep learning, gradient sparsification
@article{yoon2023sage,
title = {SAGE: toward on-the-fly gradient compression ratio scaling},
author = {Daegun Yoon and Minjoong Jeong and Sangyoon Oh},
url = {https://link.springer.com/article/10.1007/s11227-023-05120-7},
doi = {https://doi.org/10.1007/s11227-023-05120-7},
year = {2023},
date = {2023-02-25},
urldate = {2023-02-25},
journal = {The Journal of Supercomputing},
pages = {1--23},
abstract = {Gradient sparsification is widely adopted in distributed training; however, it suffers from a trade-off between computation and communication. The prevalent Top-k sparsifier has a hard constraint on computational overhead while achieving the desired gradient compression ratio. Conversely, the hard-threshold sparsifier eliminates computational constraints but fail to achieve the targeted compression ratio. Motivated by this tradeoff, we designed a novel threshold-based sparsifier called SAGE, which achieves a compression ratio close to that of the Top-k sparsifier with negligible computational overhead. SAGE scales the compression ratio by deriving an adjustable threshold based on each iteration’s heuristics. Experimental results show that SAGE achieves a compression ratio closer to the desired ratio than a hard-threshold sparsifier without exacerbating the accuracy of model training. In terms of computation time for gradient selection, SAGE achieves a speedup of up to 23.62×
over the Top-k sparsifier.},
keywords = {distributed deep learning, gradient sparsification},
pubstate = {published},
tppubtype = {article}
}
over the Top-k sparsifier.
2022
Yoon, Daegun; Jeong, Minjoong; Oh, Sangyoon
WAVE: designing a heuristics-based three-way breadth-first search on GPUs🌏 InternationalJournal Article
In: The Journal of Supercomputing, 2022, (2).
Abstract | Links | BibTeX | 태그: breath-first search, direction-optimizing BFS, GPU, graph, push-pull
@article{Yoon2022WAVE,
title = {WAVE: designing a heuristics-based three-way breadth-first search on GPUs},
author = {Daegun Yoon and Minjoong Jeong and Sangyoon Oh},
doi = {10.1007/s11227-022-04934-1},
year = {2022},
date = {2022-11-17},
urldate = {2022-11-17},
journal = {The Journal of Supercomputing},
abstract = {Breadth-first search (BFS) is a building block for improving the performance of many iterative graph algorithms. In addition to conventional BFS (push), a novel method that traverses a graph in the reverse direction (pull) has emerged and gained popularity because of its enhanced processing performance. Several frameworks have recently used a hybrid approach known as direction-optimizing BFS, which utilizes both directions. However, these frameworks are mostly interested in optimizing the procedure in each direction, instead of designing sophisticated methods for determining the appropriate direction between push and pull at each iteration. Owing to the lack of in-depth discussion on this decision, state-of-the-art direction-optimizing BFS algorithms cannot demonstrate their comprehensive performance despite improvements in the design of each direction because they select ineffective directions at each iteration. We identified that the current frameworks suffer from high computational overheads for their decisions and make decisions that are overfitted to several graph datasets used for tuning their direction selection process. Based on observations from state-of-the-art limitations, we designed a direction-optimizing method for BFS called WAVE. WAVE minimizes the computational overhead to near zero and makes more appropriate direction selection decisions than the state-of-the-art heuristics based on the characteristics extracted from the input graph dataset. In our experiments on 20 graph benchmarks, WAVE achieved speedups of up to 4.95×, 5.79×, 46.49×, and 149.67× over Enterprise, Gunrock, Tigr, and CuSha, respectively.},
note = {2},
keywords = {breath-first search, direction-optimizing BFS, GPU, graph, push-pull},
pubstate = {published},
tppubtype = {article}
}
백민석,; 오상윤,
하둡 맵리듀스와 페이지 랭크를 이용한 서울시 대중 교통 인구 이동 분석 🇰🇷 DomesticConference 📃 In press
추계학술대회 Annual Conference of KIPS (ACK 2022), 한국정보처리학회 2022, (우수논문상).
Links | BibTeX | 태그: graph, Hadoop MapReduce
@conference{백민석2022하둡,
title = {하둡 맵리듀스와 페이지 랭크를 이용한 서울시 대중 교통 인구 이동 분석 },
author = {백민석 and 오상윤},
url = {https://kiss.kstudy.com/Detail/Ar?key=3988407},
year = {2022},
date = {2022-11-04},
urldate = {2022-11-04},
booktitle = {추계학술대회 Annual Conference of KIPS (ACK 2022)},
organization = {한국정보처리학회 },
note = {우수논문상},
keywords = {graph, Hadoop MapReduce},
pubstate = {published},
tppubtype = {conference}
}
여상호,; 배민호,; 정민중,; 권오경,; 오상윤,
Crossover-SGD: A gossip-based communication in distributed deep learning for alleviating large mini-batch problem and enhancing scalability🌏 InternationalJournal Article
In: Concurrency and Computation: Practice and Experience, 2022.
Abstract | Links | BibTeX | 태그: deep learning, distributed deep learning
@article{여상호2022Crossover-SGD,
title = {Crossover-SGD: A gossip-based communication in distributed deep learning for alleviating large mini-batch problem and enhancing scalability},
author = {여상호 and 배민호 and 정민중 and 권오경 and 오상윤},
url = {https://arxiv.org/abs/2012.15198},
doi = {10.48550/arXiv.2012.15198},
year = {2022},
date = {2022-11-01},
urldate = {2022-11-01},
journal = {Concurrency and Computation: Practice and Experience},
abstract = { Distributed deep learning is an effective way to reduce the training time of deep learning for large datasets as well as complex models. However, the limited scalability caused by network overheads makes it difficult to synchronize the parameters of all workers. To resolve this problem, gossip-based methods that demonstrates stable scalability regardless of the number of workers have been proposed. However, to use gossip-based methods in general cases, the validation accuracy for a large mini-batch needs to be verified. To verify this, we first empirically study the characteristics of gossip methods in a large mini-batch problem and observe that the gossip methods preserve higher validation accuracy than AllReduce-SGD(Stochastic Gradient Descent) when the number of batch sizes is increased and the number of workers is fixed. However, the delayed parameter propagation of the gossip-based models decreases validation accuracy in large node scales. To cope with this problem, we propose Crossover-SGD that alleviates the delay propagation of weight parameters via segment-wise communication and load balancing random network topology. We also adapt hierarchical communication to limit the number of workers in gossip-based communication methods. To validate the effectiveness of our proposed method, we conduct empirical experiments and observe that our Crossover-SGD shows higher node scalability than SGP(Stochastic Gradient Push). },
keywords = {deep learning, distributed deep learning},
pubstate = {published},
tppubtype = {article}
}
윤대건,; 노병희,; 오상윤,
전술망의 라우팅 성능 개선을 위한 성능 지표 분석 기반 정책 엔진 설계🇰🇷 DomesticJournal Article
In: 한국통신학회 논문지, vol. 47, iss. 9, no. 9, pp. 1353-1359, 2022.
Abstract | Links | BibTeX | 태그: software-defined networking
@article{윤대건2022전술망,
title = {전술망의 라우팅 성능 개선을 위한 성능 지표 분석 기반 정책 엔진 설계},
author = {윤대건 and 노병희 and 오상윤},
url = {https://www.kci.go.kr/kciportal/ci/sereArticleSearch/ciSereArtiView.kci?sereArticleSearchBean.artiId=ART002877193},
year = {2022},
date = {2022-10-31},
urldate = {2022-10-31},
journal = {한국통신학회 논문지},
volume = {47},
number = {9},
issue = {9},
pages = {1353-1359},
abstract = {컴퓨팅 관련 기술 발달에 따라 군 작전 수행에서 발생하는 데이터의 규모가 매우 커지고 있으며, 이에 따라 이를 처리하기 위한 군 전술망의 성능 향상에 대한 요구 또한 점점 늘어나고 있다. 군 전술망의 특성 상 다양한 장비로 구성된 네트워크를 활용해야 하며, 이러한 상황에서 민간에서 활발히 적용되는 Software-Defined Network (SDN) 기술을 적용한다면 장비를 제공 벤더로부터 자유로운 손쉬운 네트워크 관리가 가능하다. 본 논문에서는SDN 기반 네트워크 환경에서 패킷 전송 성능 향상을 목적으로 하는 네트워크 정책 엔진 구조 설계를 소개한다.
정책 엔진은 Flow table의 Flow들이 나타내는 라우팅 경로를 수정하도록 하는 알고리즘을 포함하며 성능 개선 여부는 본 연구에서 정의한 종합 성능 지표를 통해 판단한다. 추후 본 연구에서 제안하는 전술망 라우팅 성능 개량을 위한 성능 지표 분석 기반 정책 엔진 기반의 소프트웨어를 실제 네트워크 운용 상황에 적용하고, 네트워크 성능 향상을 검증하도록 할 계획이다.},
keywords = {software-defined networking},
pubstate = {published},
tppubtype = {article}
}
정책 엔진은 Flow table의 Flow들이 나타내는 라우팅 경로를 수정하도록 하는 알고리즘을 포함하며 성능 개선 여부는 본 연구에서 정의한 종합 성능 지표를 통해 판단한다. 추후 본 연구에서 제안하는 전술망 라우팅 성능 개량을 위한 성능 지표 분석 기반 정책 엔진 기반의 소프트웨어를 실제 네트워크 운용 상황에 적용하고, 네트워크 성능 향상을 검증하도록 할 계획이다.
Lee, Seungjun; Jeong, Minjoong; Oh, Sangyoon
Is Ant Colony System better than FFD for VM placement in a heterogeneous cluster?🌏 InternationalConference
2022 IEEE International Conference on Cloud Engineering (IC2E), 2022, ISBN: 978-1-6654-9116-7.
Abstract | Links | BibTeX | 태그: ant colony system, cloud computing, resource efficiency, VM placement
@conference{Seungjun2022Ant,
title = {Is Ant Colony System better than FFD for VM placement in a heterogeneous cluster?},
author = {Seungjun Lee and Minjoong Jeong and Sangyoon Oh},
url = {https://ieeexplore.ieee.org/document/9946320},
doi = {10.1109/IC2E55432.2022.00038},
isbn = {978-1-6654-9116-7},
year = {2022},
date = {2022-09-22},
urldate = {2022-06-11},
booktitle = {2022 IEEE International Conference on Cloud Engineering (IC2E)},
pages = {277-278},
abstract = {First fit decreasing (FFD) is the most popular heuristic for virtual machine (VM) placement problems. However, FFD does not perform well in a heterogeneous cluster environment in which physical machines have different capacities. Moreover, FFD and other heuristics, such as best fit decreasing (BFD), do not effectively handle the VM placement problem if multiple resources are considered together. In this study, we analyze the reason why the ant colony system performs better than FFD for VM placement in a heterogeneous cluster. We verified our logical observations through experimental comparisons with other heuristics.},
keywords = {ant colony system, cloud computing, resource efficiency, VM placement},
pubstate = {published},
tppubtype = {conference}
}