Pular para o conteúdo principal

Escrita de artigo para BDR Special issue on Benchmarking, Performance Tuning and Optimization for Big Data Analytics

Desde o dia 20/05 estou participando a elaboração de um artigo para essa revista Big Data Research  (BDR) com pesquisadores de uma universidade de Aveiro - Portugal.

O foco da escrita no momento está em técnicas de tuning e self-tuning em banco de dados tradicionais e distribuídos, incluindo os NoSQL.

Abaixo a chamada para o artigo e o prazo foi estendido para 08/06/2020.

Big Data Research Special Issue on "Benchmarking, Performance Tuning and Optimization for Big Data Analytics"
https://www.journals.elsevier.com/big-data-research/call-for-papers/benchmarking-performance-tuning-and-optimization

Scope and Aims

Users of big data are often not computer scientists. On the other hand, it is nontrivial for even experts to optimize performance of big data applications because there are so many decisions to make. For example, users have to first choose from many different big data systems and optimization algorithms to deal with complex structured data, graph data, and streaming data. In particular, there are numerous parameters to tune to optimize performance of a specific system and it is often possible to further optimize the algorithms previously written for "small data" in order to effectively adapt them in a big data environment. To make things more complex, users may worry about not only computational running time, storage cost and response time or throughput, but also quality of results, monetary cost, security and privacy, and energy efficiency. In more traditional algorithms and relational databases, these complexities are handled by query optimizer and other automatic tuning tools (e.g ., index selection tools) and there are benchmarks to compare performance of different products and optimization algorithms. Such tools are not available for big data environment and the problem is more complicated than the problem for traditional relational databases.

Topics of Interest

We invite authors from academia and industry to submit their original research as well as review articles to present latest progresses for current development or future goals in this field.
Topics of interest include, but are not limited to:

* Theoretical and empirical performance model for big data applications
* Optimization for Machine Learning and Data Mining in big data
* Benchmark and comparative studies for big data processing and analytic platforms
* Monitoring, analysis, and visualization of performance in big data environment
* Workflow/process management & optimization in big data environment
* Performance tuning and optimization for specific big data platforms or applications (e.g., No-SQL databases, graph processing systems, stream systems, SQL-on-Hadoop databases)
* Performance tuning and optimization for specific data sets (e.g., scientific data, spatial data, temporal data, text data, images, videos, mixed datasets)
* Case studies and best practices for performance tuning for big data
* Cost model and performance prediction in big data environment
* Impact of security/privacy settings on performance of big data systems
* Self adaptive or automatic tuning tools for big data applications
* Big data application optimization on High Performance Computing (HPC) and Cloud environments

Special Issue Editors
* Zhiyuan Chen, University of Maryland, Baltimore County, U.S.A, zhchen@umbc.eduhttps://userpages.umbc.edu/~zhchen/
* Jianwu Wang, University of Maryland, Baltimore County, jianwu@umbc.eduhttp://userpages.umbc.edu/~jianwu/
* Yiming Ying, University at Albany-SUNY, U.S.A, yying@albany.eduhttps://www.albany.edu/~yy298919/
* Feng Chen, University of Texas at Dallas, U.S.A, feng.chen@utdallas.edu

Important Dates
Submission Deadline: June, 1st, 2020
Author Notification: July, 15th, 2020
Revised Manuscript Due: September, 1st, 2020
Notification of Acceptance: October, 1st, 2020
Final Manuscript Due: November, 1st, 2020
Tentative Publication Date: December, 1st, 2020

About the journal:

Big Data Research is a prestigious journal in the area of big data. It has an impact factor of 2.95. More details can be found at https://www.journals.elsevier.com/big-data-research

Paper Submission Guidelines

This issue will contain papers directly submitted from the wider research community along with selected papers from the called this year and past years' IEEE International Workshop on Benchmarking, Performance Tuning and Optimization for Big Data Applications (BPOD, https://userpages.umbc.edu/~jianwu/BPOD/). Papers selected from BPOD will be substantially extended with at least 30% difference from its conference version. All submissions will go through a two-round peer-review process by at least three international researchers.

Before submitting your manuscript, please ensure you have carefully read the Instructions for Authors for the journal. https://www.elsevier.com/journals/big-data-research/2214-5796/guide-for-authors

The complete manuscript should be submitted through the journal's submission system. https://www.editorialmanager.com/bdr/default.aspx

To ensure that you submit to the correct special issue, please select the appropriate section in the dropdown menu upon submission and choose by Article Type: VSI:BigDataBench&Optimizatio. In your cover letter, please also clearly mention the title of the SI.

Comentários

  1. O artigo foi submetido em 08/06/2020 e em 13/09/2020 obtivemos o retorno que o artigo deveria ser revisado para publicação.
    Em 13/10/2020 a nova versão foi submetida e estamos aguardando a resposta final.

    ResponderExcluir

Postar um comentário

Sinta-se a vontade para comentar. Críticas construtivas são sempre bem vindas.

Postagens mais visitadas deste blog

Connected Papers: Uma abordagem alternativa para revisão da literatura

Durante um projeto de pesquisa podemos encontrar um artigo que nos identificamos em termos de problema de pesquisa e também de solução. Então surge a vontade de saber como essa área de pesquisa se desenvolveu até chegar a esse ponto ou quais desdobramentos ocorreram a partir dessa solução proposta para identificar o estado da arte nesse tema. Podemos seguir duas abordagens:  realizar uma revisão sistemática usando palavras chaves que melhor caracterizam o tema em bibliotecas digitais de referência para encontrar artigos relacionados ou realizar snowballing ancorado nesse artigo que identificamos previamente, explorando os artigos citados (backward) ou os artigos que o citam (forward)  Mas a ferramenta Connected Papers propõe uma abordagem alternativa para essa busca. O problema inicial é dado um artigo de interesse, precisamos encontrar outros artigos relacionados de "certa forma". Find different methods and approaches to the same subject Track down the state of the art rese...

Knowledge Graphs as a source of trust for LLM-powered enterprise question answering - Leitura de Artigo

J. Sequeda, D. Allemang and B. Jacob, Knowledge Graphs as a source of trust for LLM-powered enterprise question answering, Web Semantics: Science, Services and Agents on the World Wide Web (2025), doi: https://doi.org/10.1016/j.websem.2024.100858. 1. Introduction These question answering systems that enable to chat with your structured data hold tremendous potential for transforming the way self service and data-driven decision making is executed within enterprises. Self service and data-driven decision making in organizations today is largly made through Business Intelligence (BI) and analytics reporting. Data teams gather the original data, integrate the data, build a SQL data warehouse (i.e. star schemas), and create BI dashboards and reports that are then used by business users and analysts to answer specific questions (i.e. metrics, KPIs) and make decisions. The bottleneck of this approach is that business users are only able to answer questions given the views of existing dashboa...

Knowledge Graph Toolkit (KGTK)

https://kgtk.readthedocs.io/en/latest/ KGTK represents KGs using TSV files with 4 columns labeled id, node1, label and node2. The id column is a symbol representing an identifier of an edge, corresponding to the orange circles in the diagram above. node1 represents the source of the edge, node2 represents the destination of the edge, and label represents the relation between node1 and node2. >> Quad do RDF, definir cada tripla como um grafo   KGTK defines knowledge graphs (or more generally any attributed graph or hypergraph ) as a set of nodes and a set of edges between those nodes. KGTK represents everything of meaning via an edge. Edges themselves can be attributed by having edges asserted about them, thus, KGTK can in fact represent arbitrary hypergraphs. KGTK intentionally does not distinguish attributes or qualifiers on nodes and edges from full-fledged edges, tools operating on KGTK graphs can instead interpret edges differently if they so desire. In KGTK, e...