Doutorado em Ciência da Computação

URI Permanente para esta coleção

Navegar

Submissões Recentes

Agora exibindo 1 - 20 de 25
  • Item
    Abordagem de seleção de características baseada em AUC com estimativa de probabilidade combinada a técnica de suavização de La Place
    (Universidade Federal de Goiás, 2023-09-28) Ribeiro, Guilherme Alberto Sousa; Costa, Nattane Luíza da; http://lattes.cnpq.br/9968129748669015; Barbosa, Rommel Melgaço; http://lattes.cnpq.br/6228227125338610; Barbosa, Rommel Melgaço; Lima, Marcio Dias de; Oliveira, Alexandre César Muniz de; Gonçalves, Christiane; Rodrigues, Diego de Castro
    The high dimensionality of many datasets has led to the need for dimensionality reduction algorithms that increase performance, reduce computational effort and simplify data processing in applications focused on machine learning or pattern recognition. Due to the need and importance of reduced data, this paper proposes an investigation of feature selection methods, focusing on methods that use AUC (Area Under the ROC curve). Trends in the use of feature selection methods in general and for methods using AUC as an estimator, applied to microarray data, were evaluated. A new feature selection algorithm, the AUC-based feature selection method with probability estimation and the La PLace smoothing method (AUC-EPS), was then developed. The proposed method calculates the AUC considering all possible values of each feature associated with estimation probability and the LaPlace smoothing method. Experiments were conducted to compare the proposed technique with the FAST (Feature Assessment by Sliding Thresholds) and ARCO (AUC and Rank Correlation coefficient Optimization) algorithms. Eight datasets related to gene expression in microarrays were used, all of which were used for the cross-validation experiment and four for the bootstrap experiment. The results showed that the proposed method helped improve the performance of some classifiers and in most cases with a completely different set of features than the other techniques, with some of these features identified by AUC-EPS being critical for disease identification. The work concluded that the proposed method, called AUC-EPS, selects features different from the algorithms FAST and ARCO that help to improve the performance of some classifiers and identify features that are crucial for discriminating cancer.
  • Item
    Acelerando florestas de decisão paralelas em processadores gráficos para a classificação de texto
    (Universidade Federal de Goiás, 2022-09-12) Pires, Julio Cesar Batista; Martins, Wellington Santos; http://lattes.cnpq.br/3041686206689904; Martins, Wellington Santos; Lima, Junio César de; Gaioso, Roussian Di Ramos Alves; Franco, Ricardo Augusto Pereira; Soares, Fabrízzio Alphonsus Alves de Melo Nunes
    The amount of readily available on-line text has grown exponentially, requiring efficient methods to automatically manage and sort data. Automatic text classification provides means to organize this data by associating documents with classes. However, the use of more data and sophisticated machine learning algorithms has demanded an increasingly computing power. In this work we accelerate a novel Random Forest-based classifier that has been shown to outperform state-of-art classifiers for textual data. The classifier is obtained by applying the boosting technique in bags of extremely randomized trees (forests) that are built in parallel to improve performance. Experimental results using standard textual datasets show that the GPUbased implementation is able to reduce the execution time by up to 20 times compared to an equivalent sequential implementation.
  • Item
    Abordagem de seleção de características baseada em AUC com estimativa de probabilidade combinada a técnica de suavização de La Place
    (Universidade Federal de Goiás, 2024-09-28) Ribeiro, Guilherme Alberto Sousa; Costa, Nattane Luíza da; http://lattes.cnpq.br/9968129748669015; Barbosa, Rommel Melgaço; http://lattes.cnpq.br/6228227125338610; Barbosa, Rommel Melgaço; Lima, Marcio Dias de; Oliveira, Alexandre César Muniz de; Gonçalves, Christiane; Rodrigues, Diego de Castro
    The high dimensionality of many datasets has led to the need for dimensionality reduction algorithms that increase performance, reduce computational effort and simplify data processing in applications focused on machine learning or pattern recognition. Due to the need and importance of reduced data, this paper proposes an investigation of feature selection methods, focusing on methods that use AUC (Area Under the ROC curve). Trends in the use of feature selection methods in general and for methods using AUC as an estimator, applied to microarray data, were evaluated. A new feature selection algorithm, the AUC-based feature selection method with probability estimation and the La PLace smoothing method (AUC-EPS), was then developed. The proposed method calculates the AUC considering all possible values of each feature associated with estimation probability and the La Place smoothing method. Experiments were conducted to compare the proposed technique with the FAST (Feature Assessment by Sliding Thresholds) and ARCO (AUC and Rank Correlation coefficient Optimization) algorithms. Eight datasets related to gene expression in microarrays were used, all of which were used for the crossvalidation experiment and four for the bootstrap experiment. The results showed that the proposed method helped improve the performance of some classifiers and in most cases with a completely different set of features than the other techniques, with some of these features identified by AUC-EPS being critical for disease identification. The work concluded that the proposed method, called AUC-EPS, selects features different from the algorithms FAST and ARCO that help to improve the performance of some classifiers and identify features that are crucial for discriminating cancer.
  • Item
    Análise multirresolução de imagens gigapixel para detecção de faces e pedestres
    (Universidade Federal de Goiás, 2023-09-27) Ferreira, Cristiane Bastos Rocha; Pedrini, Hélio; http://lattes.cnpq.br/9600140904712115; Soares, Fabrízzio Alphonsus Alves de Melo Nunes; http://lattes.cnpq.br/7206645857721831; Soares, Fabrízzio Alphonsus Alves de Melo Nunes; Pedrini, Helio; Santos, Edimilson Batista dos; Borges, Díbio Leandro; Fernandes, Deborah Silva Alves
    Gigapixel images, also known as gigaimages, can be formed by merging a sequence of individual images obtained from a scene scanning process. Such images can be understood as a mosaic construction based on a large number of high resolution digital images. A gigapixel image provides a powerful way to observe minimal details that are very far from the observer, allowing the development of research in many areas such as pedestrian detection, surveillance, security, and so forth. As this image category has a high volume of data captured in a sequential way, its generation is associated with many problems caused by the process of generating and analyzing them, thus, applying conventional algorithms designed for non-gigapixel images in a direct way can become unfeasible in this context. Thus, this work proposes a method for scanning, manipulating and analyzing multiresolution Gigapixel images for pedestrian and face identification applications using traditional algorithms. This approach is analyzed using both Gigapixel images with low and high density of people and faces, presenting promising results.
  • Item
    Aplicação de técnicas de visualização de informações para os problemas de agendamento de horários educacionais
    (Universidade Federal de Goiás, 2023-10-20) Alencar, Wanderley de Souza; Jradi, Walid Abdala Rfaei; http://lattes.cnpq.br/6868170610194494; Nascimento, Hugo Alexandre Dantas do; http://lattes.cnpq.br/2920005922426876; Nascimento, Hugo Alexandre Dantas do; Jradi, Walid Abdala Rfaei; Bueno, Elivelton Ferreira; Gondim, Halley Wesley Alexandre Silva; Carvalho, Cedric Luiz de
    An important category, or class, of combinatorial optimization problems is called Educational Timetabling Problems (Ed-TTPs). Broadly, this category includes problems in which it is necessary to allocate teachers, subjects (lectures) and, eventually, rooms in order to build a timetable, of classes or examinations, to be used in a certain academic period in an educational institution (school, college, university, etc.). The timetable to be prepared must observe a set of constraints in order to satisfy, as much as possible, a set of desirable goals. The current research proposes the use of methods and/or techniques from the Information Visualization (IV) area to, in an interactive approach, help a better understanding and resolution, by non-technical users, of problem instances in the scope of their educational institutions. In the proposed approach, human actions and others performed by a computational system interact in a symbiotic way targeting the problem resolution, with the interaction carried out through a graphical user interface that implements ideas originating from the User Hints framework [Nas03]. Among the main contributions achieved are: (1) recognition, and characterization, of the most used techniques for the presentation and/or visualization of Ed-TTPs solutions; (2) conception of a mathematical notation to formalize the problem specification, including the introduction of a new idea called flexibility applied to the entities involved in the timetable; (3) proposition of visualizations able to contribute to a better understanding of a problem instance; (4) make available a computational tool that provides interactive resolution of Ed-TTPs, together with a specific entity-relationship model for this kind of problem; and, finally, (5) the proposal of a methodology to evaluate visualizations applied to the problem in focus.
  • Item
    Reconhecimento de padrões em imagens radiográficas de tórax: apoiando o diagnóstico de doenças pulmonares infecciosas
    (Universidade Federal de Goiás, 2023-09-29) Fonseca, Afonso Ueslei da; Soares, Fabrízzio Alphonsus Alves de Melo Nunes; http://lattes.cnpq.br/7206645857721831; Soares, Fabrízzio Alphonsus Alves de Melo Nunes; Laureano, Gustavo Teodoro; Pedrini, Hélio; Rabahi, Marcelo Fouad; Salvini, Rogerio Lopes
    Pattern Recognition (PR) is a field of computer science that aims to develop techniques and algorithms capable of identifying regularities in complex data, enabling intelligent systems to perform complicated tasks with precision. In the context of diseases, PR plays a crucial role in diagnosis and detection, revealing patterns hidden from human eyes, assisting doctors in making decisions and identifying correlations. Infectious pulmonary diseases (IPD), such as pneumonia, tuberculosis, and COVID-19, challenge global public health, causing thousands of deaths annually, affecting healthcare systems, and demanding substantial financial resources. Diagnosing them can be challenging due to the vagueness of symptoms, similarities with other conditions, and subjectivity in clinical assessment. For instance, chest X-ray (CXR) examinations are a tedious and specialized process with significant variation among observers, leading to failures and delays in diagnosis and treatment, especially in underdeveloped countries with a scarcity of radiologists. In this thesis, we investigate PR and Artificial Intelligence (AI) techniques to support the diagnosis of IPID in CXRs. We follow the guidelines of the World Health Organization (WHO) to support the goals of the 2030 Agenda, which includes combating infectious diseases. The research questions involve selecting the best techniques, acquiring data, and creating intelligent models. As objectives, we propose low-cost, high-efficiency, and effective PR and AI methods that range from preprocessing to supporting the diagnosis of IPD in CXRs. The results so far align with the state of the art, and we believe they can contribute to the development of computer-assisted IPD diagnostic systems.
  • Item
    Recomendação de conteúdo ciente de recursos como estratégia para cache na borda da rede em sistemas 5G
    (Universidade Federal de Goiás, 2023-10-03) Monção, Ana Claudia Bastos Loureiro; Corrêa, Sand Luz; http://lattes.cnpq.br/3386409577930822; Cardoso, Kleber Vieira; http://lattes.cnpq.br/0268732896111424; Cardoso, Kleber Vieira; Corrêa, Sand Luz; Soares, Telma Woerle de Lima; Rosa, Thierson Couto; Fonseca, Anelise Munaretto
    Recently, the coupling between content caching at the wireless network edge and video recommendation systems has shown promising results to optimize the cache hit and improve the user experience. However, the quality of the UE wireless link and the resource capabilities of the UE are aspects that impact the user experience and that have been neglected in the literature. In this work, we present a resource-aware optimization model for the joint task of caching and recommending videos to mobile users. We also present a heuristic created to solve the problem more quickly. The goal is to maximize the cache hit ratio and the user QoE (concerning content preferences and video representations) under the constraints of UE capabilities and the availability of network resources by the time of the recommendation. We evaluate our proposed model using a video catalog derived from a real-world video content dataset (from the MovieLens project), real- world video representations and actual historical records of Channel Quality Indicators (CQI) representing user mobility. We compare the performance of our proposal with a state-of-the-art caching and recommendation method unaware of computing and network resources. Results show that our approach significantly increases the user’s QoE and still promotes a gain in effective cache hit rate.
  • Item
    Preditor híbrido de estruturas terciárias de proteínas
    (Universidade Federal de Goiás, 2023-08-10) Almeida, Alexandre Barbosa de; Soares, Telma Woerle de Lima; http://lattes.cnpq.br/6296363436468330; Soares , Telma Woerle de Lima; Camilo Junior , Celso Gonoalves; Vieira, Flávio Henrique Teles; Delbem, Alexandre Cláudio Botazzo; Faccioli, Rodrigo Antônio
    Proteins are organic molecules composed of chains of amino acids and play a variety of essential biological functions in the body. The native structure of a protein is the result of the folding process of its amino acids, with their spatial orientation primarily determined by two dihedral angles (φ, ψ). This work proposes a new hybrid method for predicting the tertiary structures of proteins called hyPROT, combining techniques of Multi-objective Evolutionary Algorithm optimization (MOEA), Molecular Dynamics, and Recurrent Neural Networks (RNNs). The proposed approach investigates the evolutionary profile of dihedral angles (φ, ψ) obtained by different MOEAs during the minimization process of the objective function by dominance and energy minimization by molecular dynamics. This proposal is unprecedented in the protein prediction literature. The premise under investigation is that the evolutionary profile of dihedrals may be concealing relevant patterns about folding mechanisms. To analyze the evolutionary profile of angles (φ, ψ), RNNs were used to abstract and generalize the specific biases of each MOEA. The selected MOEAs were NSGAII, BRKGA, and GDE3, and the objective function investigated combines the potential energy from non-covalent interactions and the solvation energy. The results obtained show that the hyPROT was able to reduce the RMSD value of the best prediction generated by the MOEAs individually by at least 33%. Predicting new series for dihedral angles allowed for the formation of histograms, indicating the formation of a possible statistical ensemble responsible for the distribution of dihedrals (φ, ψ) during the folding process
  • Item
    Avaliação da qualidade da sintetização de fala gerada por modelos de redes neurais profundas
    (Universidade Federal de Goiás, 2023-05-26) Oliveira, Frederico Santos de; Soares, Anderson da Silva; http://lattes.cnpq.br/1096941114079527; Soares, Anderson da Silva; Aluisio, Sandra Maria; Duarte, Julio Cesar; Laureano, Gustavo Teodoro; Galvão Filho, Arlindo Rodrigues
    With the emergence of intelligent personal assistants, the need for high-quality conversational interfaces has increased. While text-based chatbots are popular, the development of voice interfaces is equally important. However, the primary method for evaluating voice-based conversational models is mainly done through Mean Opinion Score (MOS), which relies on a manual and subjective process. In this context, this thesis aims to contribute with a new methodology for evaluating voice-based conversational interfaces, with a case study specifically conducted in Brazilian Portuguese. The proposed methodology includes an architecture for predicting the quality of synthesized speech in Brazilian Portuguese, correlated with MOS. To evaluate the proposed methodology, this work included training Text-to-Speech models to create the dataset called BRSpeechMOS. Details about the creation of this dataset are presented, along with a qualitative and quantitative analysis of it. A series of experiments were conducted to train various architectures using the BRSpeechMOS dataset. The architectures used are based on supervised and self-supervised learning. The results obtained confirm the hypothesis raised that pre-trained models on voice processing tasks such as speaker verification and automatic speech recognition produce suitable acoustic representations for the task of predicting speech quality, contributing to the advancement of the state of the art in the development of evaluation methodologies for conversational models.
  • Item
    Controle de admissão para network slicing considerando recursos de comunicação e computação
    (Universidade Federal de Goiás, 2023-05-10) Lima, Henrique Valle de; Cardoso, Kleber Vieira; http://lattes.cnpq.br/0268732896111424; Corrêa, Sand Luz; http://lattes.cnpq.br/3386409577930822; Corrêa, Sand Luz; Cardoso, Kleber Vieira; Oliveira Júnior, Antônio Carlos de; Costa, Ronaldo Martins da; Both, Cristiano Bonato
    The 5G networks have enabled the application of various innovative and disruptive technologies such as Network Function Virtualization (NFV) and Software-Defined Networking (SDN). Together, these technologies act as enablers of Network Slicing (NS), transforming the way networks are operated, managed, and monetized. Through the concept of Slice-as-a-Service (SlaaS), telecommunications operators can monetize the physical and logical infrastructure by offering network slices to new customers, such as vertical industries. This thesis addresses the problem of tenant admission control using NS. We propose three admission control models for NS (MONETS-OBD, MONETS-OBS, and CAONS) that consider both communication and computation resources. To evaluate the proposed models, we compare the results with different classical algorithms from the literature, such as eUCB, e-greedy, and ONETS. We use data from different applications to enrich the analysis. The results indicate that the MONETS-OBD, MONETS-OBS, and CAONS heuristics perform admission control that approaches the set of ideal solutions. We achieve high efficiency with the MONETS-OBD and MONETS-OBS heuristics in controlling tenant admission, reaching acceptance rates of up to 99% in some cases. Furthermore, the CAONS heuristic, which employs penalties, not only achieves acceptance and reward rates close to the optimal solution but also significantly reduces the number of capacity violations. Lastly, the results highlight that the process of slice admission control should consider both communication and computation resources, which are scarce at the network edge. A solution that considers only communication resources can lead to incorrect and unfeasible interpretations, overestimating the capacity of computation resources.
  • Item
    CLAT: arcabouço conceitual e ferramenta de apoio à avaliação da escrita inicial infantil por meio de dispositivos móveis
    (Universidade Federal de Goiás, 2022-12-21) Mombach, Jaline Gonçalves; Soares, Fabrizzio Alphonsus Alves de Melo Nunes; http://lattes.cnpq.br/7206645857721831; Soares, Fabrizzio Alphonsus Alves de Melo Nunes; Ferreira, Deller James; Marques, Fátima de Lourdes dos Santos Nunes; Rodrigues, Kamila Rios da Hora; Rocha, Maria Alice de Sousa Carvalho
    In childhood literacy, the assessment of initial writing is essential for monitoring learning and consequently planning more effective interventions by educators. However, during the Covid19 pandemic period, early spelling assessments were hampered since the digital tools available did not include some strategic signals, such as visualization of the child's tracing, the reading mode, and the genuine child's thinking about writing. Therefore, as a research problem, we investigate how mobile devices could support the remote child's spelling assessment. Thus, the central goal was to develop an interaction model for mobile devices to promote these writing assignments remotely. Thus, we adopted Design Science Research as a methodological approach. In the study of the problem stage, we conducted a systematic mapping study, a survey with professionals and parents, and we documented the usability requirements. Next, we proposed an artifact for educators to create digital assignments and another to capture the children's tracing and the mode they read. Finally, for validation, we performed concept tests to teachers, children, and a validation experiment in the school ecosystem, involving 92 children and six teachers. The results indicated that children were expressively interested in the resource and could interact satisfactorily on the digital artifact, validating the interaction modeling by registering their writing without significant difficulties. Furthermore, the teachers declared that it is possible to evaluate the children's spelling from the registers visualized on the digital artifact and emphasized the similarity between the interactions promoted by artifacts and the face-to-face environment. The findings of this study contribute to research on digital writing development and new educational resources. At the social level, the proposal also contributes directly to the maintenance of teaching in remote environments while also bringing new possibilities for face-to-face teaching and blended learning.
  • Item
    Atribuição de papéis em alguns produtos de grafos
    (Universidade Federal de Goiás, 2022-06-24) Mesquita, Fernanda Neiva; Dias, Elisângela Silva; http://lattes.cnpq.br/0138908377103572; Nascimento, Julliano Rosa; http://lattes.cnpq.br/8971175373328824; Castonguay, Diane; http://lattes.cnpq.br/4005898623592261; Castonguay, Diane; Rodrigues, Rosiane de Freitas; Dourado, Mitre Costa; Nobrega, Diana Sasaki; Silva, Hebert Coelho da
    During the pandemic, due to the new coronavirus (COVID-19), the use of social networks was enhanced by social distancing and the need to stay connected, generating a gigantic volume of data. In order to extract information, graphs constitute a powerful modeling tool in which the vertices represent individuals and the edges represent relationships between them. In 1991, Everett and Borgatti formalized the concept of role assignment under the name role coloring. Thus, a r-role assignment of a simple graph G is an assignment of r distinct roles to the vertices of G, such that two vertices with the same role have the same set of roles in the related vertices. Furthermore, a specific r-role assignment defines a role graph, in which the vertices are the distinct r roles, and there is an edge between two roles whenever there are two related vertices in the graph G that correspond to these roles. Research on role assignment and operations on graphs is scarce. We showed a dichotomy for the r-role assignment problem for the Cartesian product. While the Cartesian product of two graphs always admits a 2-role assignment, the problem remains NP-complete for any fixed r ≥ 3. The complementary prism arises from the complementary product, introduced by Haynes, Henning and Van Der Merwe in 2019, which is a generalization of the Cartesian product. Complementary prisms admits a 2-role assignment, with the exception of the complementary prism of a path with three vertices. We verified that the complementary prisms admits a 3-role assignment, with the exception of the complementary prism of some not connected bipartite graphs. Next, we showed that the related problem can be solved in linear time. Finally, we conjecture that, for r ≥ 3 the problem of (r+1)-role assignment to complementary prisms is NP-complete. In this sense, we consider the role graph K'_{1,r} which is the bipartite graph K_{1,r} with a loop at the vertex of degree r and we highlight that the problem of deciding whether a prism complement has a (r+1)-role assignment, when the role graph is K'_{1,r}, it is NP-complete.
  • Item
    Geração de modelos sintéticos de topologia de sistemas de distribuição de energia elétrica
    (Universidade Federal de Goiás, 2022-05-23) Teles, Ronneesley Moura; Camillo, Marcos Henrique Marçal; http://lattes.cnpq.br/6702028822213616; Soares, Telma Woerle de Lima; http://lattes.cnpq.br/6296363436468330; Soares, Telma Woerle de Lima; Fanucchi, Rodrigo Zempulski; London Junior, João Bosco Augusto; Galvão Filho, Arlindo Rodrigues; Camillo, Marcos Henrique Marçal
    The development of algorithms to solve problems related to electrical networks has always faced access to data, since, for reasons of security and secrecy, analysts are unable to obtain real data from these networks. This work arises in order to mitigate this problem, allowing these researchers to generate synthetic networks with characteristics close to real networks. For this, algorithms were developed to generate forests with several rooted trees based on the characteristics of real electrical networks. These algorithms are guided during the search process through the desired topological distributions, allowing the use of characteristics of real networks from any Brazilian state or regions of the world. This study was based on the electrical networks of the state of Paraná. Each proposed algorithm was studied in relation to its tree generation trends allowing a better understanding of its behavior. These algorithms were employed in the generation of electrical networks through a multiobjective evolutionary computation process, using the NSGA-II method, and resulted in networks with degree distribution, number of buses per feeder and number of leaves per feeder similar to the real networks studied. In addition, a method for positioning consumers was determined and a single-objective genetic algorithm was developed for the ideal positioning of normally closed and normally open switches. It was able to evaluate faults in all the buses of an electrical network in a single depth search through the use of dynamic programming techniques. By using the algorithms and methodology developed in this thesis, the researcher will have a topology containing feeder, buses, sections, normally open and normally closed switches, and the number of consumers in the buses at his disposal for his studies.
  • Item
    Aprimoramento do modelo de seleção dos padrões associativos: uma abordagem de mineração de dados
    (Universidade Federal de Goiás, 2021-12-20) Rodrigues, Diego de Castro; Barbosa, Rommel Melgaço; http://lattes.cnpq.br/6228227125338610; Barbosa, Rommel Melgaço; Costa, Ronaldo Martins da; Costa, Nattane Luíza da; Rocha, Marcelo Lisboa; Jorge, Lúcio de Castro
    The objective of this study is to improve the association rule selection model through a set of asymmetric probabilistic metrics. We present the Health Association Rules - HAR, based on Apriori, the algorithm is composed of six functions and uses alternative metrics to the Support/Confidence model to identify the implication X → Y . Initially, the application of our solution was focused only on health data, but we realized that asymmetrical associative patterns could be applied in other contexts that seek to address the cause and effect of a pattern. Our experiments were composed of 60 real datasets taken from specialist websites, research partnerships and open data. We empirically observed the behavior of HAR in all data sets, and a comparison was performed with the classical Apriori algorithm. We realized that it has overcome the main problems of the Support/Confidence model. We were able to identify the most relevant patterns for the observed datasets, eliminating logical contradictions and redundancies. We also perform a statistical analysis of the experiments where the statistical effect is positive for HAR. HAR was able to discover more representative patterns and rare patterns, in addition to being able to perform rule grouping, filtering and ranking. Our solution presented a linear behavior in the experiments, being able to be applied in health, social, content suggestion, product indication and educational data. Not limited to these data domains, HAR is prepared to receive large amounts of data by using a customized parallel architecture.
  • Item
    Detecção automática e avaliação de linhas de plantio de cana-de-açúcar em imagens aéreas
    (Universidade Federal de Goiás, 2021-12-09) Rocha, Bruno Moraes; Pedrini, Hélio; http://lattes.cnpq.br/9600140904712115 Nome completo do 2º coorientador(a): E-mail: Nomes completos; Soares, Fabrízzio Alphonsus Alves de Melo Nunes; http://lattes.cnpq.br/7206645857721831; Soares, Fabrízzio Alphonsus Alves de Melo Nunes; Pedrini, Hélio; Salvini, Rogerio Lopes; Costa, Ronaldo Martins da; Cabacinha, Christian Dias
    For higher productivity and economic yield in sugarcane field, several imaging techniques using sugarcane field images have been developed. However, the identification and measurement of gaps in sugarcane field crop rows are still commonly performed manually on site to decide to replant the gaps or the entire area. Manual measurement has a high cost of time and manpower. Based on these factors, this study aimed to create a new technique that automatically identifies and evaluates the gaps along the crop rows in aerial images of sugarcane fields obtained by a small remotely piloted aircraft. The images captured using the remotely piloted aircraft were used to generate the orthomosaics of the crop field area and classified with the algorithm K-Nearest Neighbors to segment the crop rows. The orientation of the planting rows in the image was found using the filter gradient Red Green Blue. Then, the crop rows were mapped using the curve adjustment method and overlap the classified image to detect and measure the gaps along the segment of the planting line. The technique developed obtained a maximum error of approximately 3% when compared to the manual method to evaluate the length of the gaps in the crop rows in an orthomosaic with an area of 8.05 hectares using the method proposed by Stolf, adapted for digital images. The proposed approach was able to properly identify the spatial position of automatically generated line segments over manually created line segments. The proposed method was also able to achieve statistically similar results when confronted with the technique performed manually in the image for the mapping of rows and identification of gaps for sugarcane fields with growth 40 and 80 days after planting. The automatic technique developed had a significant result in the evaluation of the gaps in the crop rows in the aerial images of sugarcane fields, thus, its use allows automated inspections with high accuracy measurements, and besides being able to assist producers in making decisions in the management of their sugarcane fields.
  • Item
    Escalonamento de recursos em redes sem fio 5G baseado em otimização de retardo e de alocação de potência considerando comunicação dispositivo a dispositivo
    (Universidade Federal de Goiás, 2021-10-15) Ferreira, Marcus Vinícius Gonzaga; Vieira, Flávio Henrique Teles; http://lattes.cnpq.br/0920629723928382; Vieira, Flávio Henrique Teles; Madeira, Edmundo Roberto Mauro; Lima, Marcos Antônio Cardoso de; Rocha, Flávio Geraldo Coelho; Oliveira Júnior, Antônio Carlos de
    In this thesis, a resources scheduling scheme is proposed for 5G wireless network based on CP-OFDM (Cyclic Prefix - Orthogonal Frequency Division Multiplexing) and f-OFDM (filtered - OFDM) modulations in order to optimize the average delay and the power allocation for users. In the proposed approach the transmission rate value is calculated and the modulation format is defined so that minimize system BER (Bits Error Rate). The algorithm considers, in addition to the transmission modes determined to minimize the BER, the calculation of the system's weighted throughput to optimize the users' average delay. Additionally, it is proposed an algorithm for uplink transmission in 5G wireless networks with D2D (Device-to-device) multi-sharing communication which initially allocates resources for the CUEs (Cellular User Equipments) and subsequently allocates network resources for communication between DUEs (D2D User Equipment) pairs based in the optimization of the delay and power allocation. The proposed algorithm, namely DMCG (Delay Minimization Conflict Graph), considers the minimization of the estimated delay function using concepts of Network Calculus to decide on the allocation of idle resources of the network CUEs for DUEs pairs. In this thesis, the performance of the proposed algorithms for downlink and uplink transmission are verified and compared with others algorithms in the literature in terms of several QoS (Quality of Service) parameters and considering the carrier aggregation and 256-QAM (Quadrature Amplitude Modulation) technologies. In computational simulations they are also considered scenarios with propagation by millimeter waves and the 5G specifications of the 3GPP (3rd Generation Partnership Project) Release 15. The simulation results show that the algorithms proposed for downlink and uplink transmission provide better system performance in terms of throughput and delay, in addition to presenting lower processing time compared to optimization heuristics and other QoS parameters being compatible to those of the compared algorithms.
  • Item
    Uso e estabilidade de seletores de variáveis baseados nos pesos de conexão de redes neurais artificiais
    (Universidade Federal de Goiás, 2021-03-19) Costa, Nattane Luíza da; Barbosa, Rommel Melgaço; http://lattes.cnpq.br/6228227125338610; Barbosa, Rommel Melgaço; Lima, Márcio Dias de; Lins, Isis Didier; Costa, Ronaldo Martins da; Leitão Júnior, Plínio de Sá
    Artificial Neural Networks (ANN) are machine learning models used to solve problems in several research fields. Although, ANNs are often considered “black boxes”, which means that these models cannot be interpreted, as they do not provide explanatory information. Connection Weight Based Feature Selectors (WBFS) have been proposed to extract knowledge from ANNs. Most of studies that have been using these algorithms are based on just one ANN model. However, there are variations in the ANN connection weight values due to the initialization and training, and consequently, leading to variations in the importance ranking generated by a WBFS. In this context, this thesis presents a study about the WBFS. First, a new voting approach is proposed to assess the stability of the WBFS, i.e, the variation in the result of the WBFS. Then, we evaluated the stability of the algorithms based on multilayer perceptron (MLP) and extreme learning machines (ELM). Furthermore, an improvement is proposed in the algorithms of Garson, Olden, and Yoon, combining them with the feature selector ReliefF. The new algorithms are called FSGR, FSOR, and FSYR. The experiments were performed based on 28 MLP architectures, 16 ELM architectures, and 16 data sets from the UCI Machine Learning Repository. The results show that there is a significant difference in WBFS stability depending on the training parameters of the ANNs and depending on the WBFS used. In addition, the proposed algorithms proved to be more effective than the classic algorithms. As far as we know, this study was the first attempt to measure the stability of WBFS, to investigate the effects of different ANN training parameters on the stability of WBFS, and the first to propose a combination of WBFS with another feature selector. Besides, the results provide information about the benefits and limitations of WBFS and represent a starting point for improving the stability of these algorithms.
  • Item
    Exploiting parallelism in document similarity tasks with applications
    (Universidade Federal de Goiás, 2019-09-05) Amorim, Leonardo Afonso; Martins, Wellington Santos; http://lattes.cnpq.br/3041686206689904; Martins, Wellington Santos; Vincenzi, Auri Marcelo Rizzo; Rodrigues, Cássio Leonardo; Rosa, Thierson Couto; Martins, Weber
    The amount of data available continues to grow rapidly and much of it corresponds to text expressing human language, that is unstructured in nature. One way of giving some structure to this data is by converting the documents to a vector of features corresponding to word frequencies (term count, tf-idf, etc) or word embeddings. This transformation allows us to process textual data with operations such as similarity measure, similarity search, classification, among others. However, this is only possible thanks to more sophisticated algorithms which demand higher computational power. In this work, we exploit parallelism to enable the use of parallel algorithms to document similarity tasks and apply some of the results to an important application in software engineering. The similarity search for textual data is commonly performed through a k nearest neighbor search in which pairs of document vectors are compared and the k most similar are returned. For this task we present FaSSTkNN, a fine-grain parallel algorithm, that applies filtering techniques based on the most common important terms of the query document using tf-idf. The algorithm implemented on a GPU improved the top k nearest neighbors search by up to 60x compared to a baseline, also running on a GPU. Document similarity using tf-idf is based on a scoring scheme for words that reflects how important a word is to a document in a collection. Recently a more sophisticated similarity measure, called word embedding, has become popular. It creates a vector for each word that indicates co-occurrence relationships between words in a given context, capturing complex semantic relationships between words. In order to generate word embeddings efficiently, we propose a fine-grain parallel algorithm that finds the k less similar or farthest neighbor words to generate negative samples to create the embeddings. The algorithm implemented on a multi-GPU system scaled linearly and was able to generate embeddings 13x faster than the original multicore Word2Vec algorithm while keeping the accuracy of the results at the same level as those produced by standard word embedding programs. Finally, we applied our accelerated word embeddings solution to the problem of assessing the quality of fixes in Automated Software Repair. The proposed implementation was able to deal with large corpus, in a computationally efficient way, being a promising alternative to the processing of million source code files needed for this task.
  • Item
    Classificação de cenas utilizando a análise da aleatoriedade por aproximação da complexidade de Kolmogorov
    (Universidade Federal de Goiás, 2020-03-15) Feitosa, Rafael Divino Ferreira; Soares, Anderson da Silva; http://lattes.cnpq.br/1096941114079527; Soares, Anderson da Silva; Delbem, Alexandre Cláudio Botazzo; Soares, Fabrízzio Alphonsus Alves de Melo Nunes; Laureano, Gustavo Teodoro; Costa, Ronaldo Martins da
    In many pattern recognition problems, discriminant features are unknown and/or class boundaries are not well defined. Several studies have used data compression to discover knowledge, without features extraction and selection. The basic idea is two distinct objects can be grouped as similar, if the information content of one explains, in a significant way, the information content of the other. However, compressionbased techniques are not efficient for images, as they disregard the semantics present in the spatial correlation of two-dimensional data. A classifier is proposed for estimates the visual complexity of scenes, namely Pattern Recognition by Randomness (PRR). The operation of the method is based on data transformations, which expand the most discriminating features and suppress details. The main contribution of the work is the use of randomness as a measure discrimination. The approximation between scenes and trained models, based on representational distortion, promotes a lossy compression process. This loss is associated with irrelevant details, when the scene is reconstructed with the representation of true class, or with the information degradation, when it is reconstructed with divergent representations. The more information preserved, the greater the randomness of the reconstruction. From the mathematical point of view, the method is explained by two main measures in the U-dimensional plane: intersection and dispersion. The results yielded accuracy of 0.6967, for a 12-class problem, and 0.9286 for 7 classes. Compared with k-NN and a data mining toolkit, the proposed classifier was superior. The method is capable of generating efficient models from few training samples. It is invariant for vertical and horizontal reflections and resistant to some geometric transformations and image processing.
  • Item
    Problema de particionamento em subgrafos complementares: complexidade e convexidade
    (Universidade Federal de Goiás, 2019-11-11) Nascimento, Julliano Rosa; Coelho, Erika Morais Martins; http://lattes.cnpq.br/9389487015938509; Castonguay, Diane; http://lattes.cnpq.br/4005898623592261; Castonguay, Diane; Coelho, Erika Morais Martins; Protti, Fábio; Szwarcfiter, Jayme Luiz; Pinto, Leizer de Lima
    In this work, we introduce the PARTITION INTO COMPLEMENTARY SUBGRAPHS (COMP-SUB(Pi)) problem, which receives as input a graph H and an edge set property Pi, and the goal is determining whether is possible to decompose the graph H into complementary subgraphs G and \bar{G} such that the edge set M between G and \bar{G} satisfies property Pi. COMP-SUB(Pi) generalizes the recognition of complementary prisms problem, which is the case when Pi is a perfect matching between corresponding vertices of G and \bar{G}. When Pi is arbitrary, we show results for k-clique or k-independent set free graphs. On property P_\emptyset which considers M =\emptyset, we show that COMP-SUB(P_\emptyset) is GI-complete for chordal graphs, but can be solved efficiently for permutation, comparability, co- comparability and co-interval graphs. Furthermore, we obtain characterizations for some subclasses of chordal graphs. We also obtain results for Pi_{Kn,n} , the case when M has all the possible edges between G and \bar{G} and for Pi_{PERF}, the case which considers M as a perfect matching. In particular, we show that COMP-SUB(Pi_{PERF}) problem is GI-hard, and we obtain characterizations for this problem when the input graph H is a cograph, a chordal or a distance-hereditary graph. On the other hand, we address three parameters of the geodetic convexity for complementary prisms: the hull number, the geodetic number and the convexity number. We obtain results on the hull number for complementary prisms G\bar{G} when both G e \bar{G} are connected. On the second and third parameter, we show that the decision problems related to the geodetic number and convexity number are NP-complete even restricted to complementary prisms. We also establish lower bounds on the geodetic number for G\bar{G} when G or \bar{G} have simplicial vertices and we determine the convexity number for G\bar{G} when G is disconnected, or G is a cograph.