This document discusses recent improvements to Oracle Exadata performance, including improved SQL monitoring in Oracle 12c, enhancements to storage indexes and flash caching, and additional metrics available in AWR. It provides details on new execution plan line level metrics in SQL monitoring reports and metrics for storage cell components now visible in AWR. The post outlines various flash cache features and behavior in earlier Oracle releases.
Exadata and Database Machine Overview
The document provides an overview of Oracle's Exadata and Database Machine products. It discusses that Exadata delivers revolutionary performance that is 10-100x faster than traditional data warehouses. It then outlines the agenda and describes the Exadata architecture, features and performance capabilities. The Exadata storage servers work together in a grid configuration to deliver extreme performance for data warehousing, OLTP and consolidation workloads.
This presentation talks about the different ways of getting SQL Monitoring reports, reading them correctly, common issues with SQL Monitoring reports - and plenty of Oracle 12c-specific improvements!
Modern Linux Performance Tools for Application TroubleshootingTanel Poder
Modern Linux Performance Tools for Application Troubleshooting.
Mostly demos and focused on application/process troubleshooting, not systemwide summaries.
The document discusses techniques for analyzing SQL performance on Oracle Exadata systems using tools like ASH, SQL monitoring, and ExaSnapper. It provides examples of using these tools to identify SQL statements that are not optimized for Exadata's Smart Scan feature and determining if problematic statements are long-running queries or frequent short queries. The document also demonstrates how to selectively force full table scans for reporting workloads while keeping indexes available for OLTP workloads.
Exadata has been around since 2008 and the software features are being enhanced each release. This Presentation talks about the 12.1.x.x series of Software updates and some of the things you can now do with Exadata
Oracle Exadata is a packaged solution offering from Oracle, configured with bundled hardware, storage and database, which is touted to be optimized for handling scalable data warehouse-type workloads in query and analysis.
The document summarizes new features in Oracle Database 12c from Oracle 11g that would help a DBA currently using 11g. It lists and briefly describes features such as the READ privilege, temporary undo, online data file move, DDL logging, and many others. The objectives are to make the DBA aware of useful 12c features when working with a 12c database and to discuss each feature at a high level within 90 seconds.
This is a high level presentation I delivered at BIWA Summit. It's just some high level thoughts related to today's NoSQL and Hadoop SQL engines (not deeply technical).
The document provides a critical review of the Oracle Exadata X2-8 system. It summarizes that Exadata combines Oracle database software with specialized hardware including flash storage and servers. While Exadata promises high performance, it relies on older disk technology and is very expensive, costing over $9 million for hardware and software licenses. Alternative configurations using newer flash storage technologies from other vendors can provide similar or better performance at a lower cost without locking the customer into Oracle's platform.
The document is a presentation about Exadata and Oracle 11gR2. It provides background on the speaker's experience with parallel query over the years. It then covers key features of Exadata storage solutions and new parallel query capabilities in 11gR2 like auto degree of parallelism and statement queuing. The presentation notes some practical considerations for implementing Exadata and parallel query, such as partitioning workloads and monitoring statement queuing.
This document discusses moving data between Oracle Exadata and Hadoop for fast loading. It begins by introducing Oracle SQL Connector for HDFS, which allows querying Hadoop data using Oracle SQL and external tables. However, initial tests of loading 1TB of data showed slow speeds of only 75MB/second due to bottlenecks. Subsequent tests revealed the bottleneck was datatype conversion CPU usage on the Oracle side. The document then introduces Oracle Loader for Hadoop, which offloads datatype conversion to Hadoop cluster CPUs, allowing much faster loading of over 1GB/second by leveraging all available CPUs. Proper partitioning is also required for direct path loads to avoid contention.
Hotsos 2011: Mining the AWR repository for Capacity Planning, Visualization, ...Kristofferson A
The document discusses mining the Automatic Workload Repository (AWR) in Oracle databases for capacity planning, visualization, and other real-world uses. It introduces Karl Arao as a speaker and discusses topics he will cover including AWR, diagnosing performance issues using AWR data, visualization of AWR data, capacity planning, and tools for working with AWR data like scripts and linear regression. References and resources on working with AWR are also provided.
The document describes Oracle Exadata, a database machine that provides extreme performance for online transaction processing (OLTP) and data warehousing (DWH) workloads. It highlights key features like Smart Scans, Storage Indexes, and Flash Cache that reduce data processing loads on database servers. Customer testimonials report performance improvements like queries speeding up from days to minutes.
This document provides an overview and agenda for a presentation on tuning Oracle GoldenGate performance. It discusses measuring baseline GoldenGate performance metrics like lag times and checkpoints. It also covers tuning GoldenGate configurations like using multiple process groups. The document recommends tuning the operating system by monitoring CPU, memory, and disk I/O performance and addressing any bottlenecks found. The goal of these tuning efforts is to reduce lag times and optimize GoldenGate throughput.
Python and Oracle : allies for best of data managementLaurent Leturgez
In this presentation, I described Python and how Python can Interact with Oracle database, and Oracle Cloud Infrastructure in various project : from data visualisation to data science.
Oaktable World 2014 Kevin Closson: SLOB – For More Than I/O!Kyle Hailey
The document discusses using SLOB (Synthetic Load On Box) to test various Oracle database configurations and platforms. SLOB is described as a simple and predictable workload generator that allows testing the performance of databases under different conditions with minimal variability. The document outlines several potential uses of SLOB, including testing Oracle in-memory database options, multitenant architectures, and measuring the impact of database contention. It provides examples of using SLOB to analyze CPU and storage I/O performance.
Oracle Database 12c - The Best Oracle Database 12c Tuning Features for Develo...Alex Zaballa
Oracle Database 12c includes many new tuning features for developers and DBAs. Some key features include:
- Multitenant architecture allows multiple pluggable databases to consolidate workloads on a single database instance for improved utilization and administration.
- In-memory column store enables real-time analytics on frequently accessed data held entirely in memory for faster performance.
- New SQL syntax like FETCH FIRST for row limiting and offsetting provides more readable and intuitive replacements for previous techniques.
- Adaptive query optimization allows queries to utilize different execution plans like switching between nested loops and hash joins based on runtime statistics for improved performance.
Using VirtualBox - Learn Oracle Database 12c and EBS R12Biju Thomas
VirtualBox allows users to run multiple operating systems on a single machine. It is free to use and install. This document discusses how to install VirtualBox, import pre-built Oracle and EBS virtual machines, and find various learning resources for working with Oracle Database 12c and EBS R12 using free virtual machines and VirtualBox. Tips are provided on testing configurations in VirtualBox before moving to a production Oracle VM Server environment.
This document discusses connecting Hadoop and Oracle databases. It introduces the author Tanel Poder and his expertise in databases and big data. It then covers tools like Sqoop that can be used to load data between Hadoop and Oracle databases. It also discusses using query offloading to query Hadoop data directly from Oracle as if it were in an Oracle database.
Tanel Poder - Troubleshooting Complex Oracle Performance Issues - Part 2Tanel Poder
This document summarizes a series of performance issues seen by the author in their work with Oracle Exadata systems. It describes random session hangs occurring across several minutes, with long transaction locks and I/O waits seen. Analysis of AWR reports and blocking trees revealed that many sessions were blocked waiting on I/O, though initial I/O metrics from the OS did not show issues. Further analysis using ASH activity breakdowns and OS tools like sar and vmstat found high apparent CPU usage in ASH that was not reflected in actual low CPU load on the system. This discrepancy was due to the way ASH attributes non-waiting time to CPU. The root cause remained unclear.
Exadata and the Oracle Optimizer: The Untold StoryEnkitec
The document discusses how the Oracle optimizer can select suboptimal plans on Exadata due to not properly accounting for the faster speeds of full table scans enabled by Exadata's offloading capabilities and storage compression. It explains how updating statistics to reflect Exadata's multi-block read capabilities and gathering new system statistics can help the optimizer generate better plans that leverage Exadata's performance features.
Oracle LOB Internals and Performance TuningTanel Poder
The document discusses a presentation on tuning Oracle LOBs (Large Objects). It covers LOB architecture including inline vs out-of-line storage, LOB locators, inodes, indexes and segments. The presentation agenda includes introduction, storing large content, LOB internals, physical storage planning, caching tuning, loading LOBs, development strategies and temporary LOBs. Examples are provided to illustrate LOB structures like locators, inodes and indexes.
Oracle Latch and Mutex Contention TroubleshootingTanel Poder
This is an intro to latch & mutex contention troubleshooting which I've delivered at Hotsos Symposium, UKOUG Conference etc... It's also the starting point of my Latch & Mutex contention sections in my Advanced Oracle Troubleshooting online seminar - but we go much deeper there :-)
Oracle Database In-Memory Option in ActionTanel Poder
The document discusses Oracle Database In-Memory option and how it improves performance of data retrieval and processing queries. It provides examples of running a simple aggregation query with and without various performance features like In-Memory, vector processing and bloom filters enabled. Enabling these features reduces query elapsed time from 17 seconds to just 3 seconds by minimizing disk I/O and leveraging CPU optimizations like SIMD vector processing.
Migrating Oracle Databases to Exadata requires careful preparation to simplify and optimize databases for best performance and availability. The document discusses key points:
1. Preparation is essential to remove unnecessary objects and optimize databases before migrating.
2. Different migration methods like transportable tablespaces, data pump, or GoldenGate have advantages depending on environment and goals.
3. A fast network reduces migration time, but other bottlenecks like source system I/O or small transfers must also be addressed.
This document outlines Oracle's general product direction for informational purposes only. It does not constitute a legal commitment and should not be relied upon for purchasing decisions. The development, release, and timing of any product features described remains at Oracle's sole discretion.
The document discusses best practices for gathering statistics in Oracle databases. It covers how to gather statistics using the DBMS_STATS package, additional types of statistics like column groups and expression statistics, when to gather statistics such as after data loads, and how to improve statistics gathering performance using parallel execution and incremental gathering for partitioned tables.
Tuning SQL for Oracle Exadata: The Good, The Bad, and The Ugly Tuning SQL fo...Enkitec
This document discusses tuning SQL on Oracle Exadata. It makes three main points:
1. Gathering and displaying execution plan data differs slightly on Exadata compared to non-Exadata databases.
2. The general approach to optimization is similar to non-Exadata, focusing on features like smart scans, storage indexes, and parallelism that provide the most benefit.
3. Rewriting SQL queries can have a dramatic impact on performance by enabling offloading and reducing disk I/O, with examples showing savings of over 98% in run time.
Gluent New World #02 - SQL-on-Hadoop : A bit of History, Current State-of-the...Mark Rittman
Hadoop and NoSQL platforms initially focused on Java developers and slow but massively-scalable MapReduce jobs as an alternative to high-end but limited-scale analytics RDBMS engines. Apache Hive opened-up Hadoop to non-programmers by adding a SQL query engine and relational-style metadata layered over raw HDFS storage, and since then open-source initiatives such as Hive Stinger, Cloudera Impala and Apache Drill along with proprietary solutions from closed-source vendors have extended SQL-on-Hadoop’s capabilities into areas such as low-latency ad-hoc queries, ACID-compliant transactions and schema-less data discovery – at massive scale and with compelling economics.
In this session we’ll focus on technical foundations around SQL-on-Hadoop, first reviewing the basic platform Apache Hive provides and then looking in more detail at how ad-hoc querying, ACID-compliant transactions and data discovery engines work along with more specialised underlying storage that each now work best with – and we’ll take a look to the future to see how SQL querying, data integration and analytics are likely to come together in the next five years to make Hadoop the default platform running mixed old-world/new-world analytics workloads.
SQLcl overview - A new Command Line Interface for Oracle DatabaseJeff Smith
From the makers of Oracle SQL Developer, we present you a new take on SQL*Plus. A command line interface with a SQL history, table name completion, new commands like CTAS, DDL, Info, and simple things like editing your statement buffers using your keyboard up and down arrow keys!
Accenture Enkitec Group: Oracle database and Engineered SystemsAccenture Operations
A global leader in Oracle Exadata implementations, Accenture Enkitec Group has one of the largest concentrations of senior talent in Oracle database and Engineered Systems in the world. Our professionals are equipped with the necessary skills, plus the breadth and depth of experience to help our clients design and tune any Oracle database, and in particular to reap the full benefits of Oracle Engineered Systems
Find out more here http://bit.ly/2feNHjB
Follow us on Twitter here https://twitter.com/AccentureOps
LinkedIn https://www.linkedin.com/company/accenture-operations
If you’re already a SQL user then working with Hadoop may be a little easier than you think, thanks to Apache Hive. It provides a mechanism to project structure onto the data in Hadoop and to query that data using a SQL-like language called HiveQL (HQL).
This cheat sheet covers:
-- Query
-- Metadata
-- SQL Compatibility
-- Command Line
-- Hive Shell
The document outlines topics covered in "The Impala Cookbook" published by Cloudera. It discusses physical and schema design best practices for Impala, including recommendations for data types, partition design, file formats, and block size. It also covers estimating and managing Impala's memory usage, and how to identify the cause when queries exceed memory limits.
Geek Sync I Need for Speed: In-Memory Databases in Oracle and SQL ServerIDERA Software
You can watch the replay for this Geek Sync webcast in the IDERA Resource Center: http://ow.ly/S6MG50A5ok5
Microsoft introduced IN-MEMORY OLTP, widely referred to as “Hekaton” in SQL Server 2014. Hekaton allows for the creation of fully transactionally consistent memory-resident tables designed for high concurrency and no blocking. With SQL 2016, many of the original restrictions and limitations of this feature have been reduced. IDERA’s Vicky Harp will give an overview of this feature, including how to compile T-SQL code into machine code for an even greater performance boost.
There’s also been a lot of buzz about Oracle 12c’s new IN-MEMORY COLUMN STORE. Oracle ACE Bert Scalzo will cover this new feature, how it works, it’s benefits, scripts to measure/monitor it and more. He will also touch on performance observations from benchmarking this new feature against more traditional SGA memory allocations plus Oracle 11g R2’s Database Smart Flash Cache. All findings, scripts and conclusions from this exercise will be shared. In addition, two very popular database benchmarking tools will be highlighted.
This document discusses various MySQL performance metrics that are important to measure from within the database, operating system, and application. It outlines key InnoDB internal structures like the buffer pool and log system. Specific metrics that provide insight into buffer pool usage, page churn, and log writes are highlighted. Optimizing the working set size and ensuring sufficient free space in the log files are important factors for performance.
Caching and tuning fun for high scalabilityWim Godden
Caching has been a 'hot' topic for a few years. But caching takes more than merely taking data and putting it in a cache : the right caching techniques can improve performance and reduce load significantly. But we'll also look at some major pitfalls, showing that caching the wrong way can bring down your site. If you're looking for a clear explanation about various caching techniques and tools like Memcached, Nginx and Varnish, as well as ways to deploy them in an efficient way, this talk is for you.
This document discusses how to monitor an IBM Db2 Analytics Accelerator (IDAA). It provides an overview of the resources, use cases, and tools for monitoring an IDAA. Key metrics for monitoring include accelerator resources, system resources, SQL statements, workload, performance, and capacity planning. Tools mentioned for monitoring include the appliance UI, OMPE, Data Studio, DISPLAY ACCEL command, and stored procedures.
Trace flags are used to temporarily change SQL Server's behavior for debugging or diagnosing issues. This document discusses several trace flags including:
TF 652, 661, 834, 836 which disable certain SQL Server processes or enable large page allocations.
TF 1211, 1224 which avoid lock escalation. TF 1117 forces data files to auto grow equally. TF 1204, 1205, 1222 provide more information on deadlocks.
TF 1118 addresses tempdb contention. TFs 3226, 3014, 3004 provide more backup/restore details. TF 4199 enables query processor fixes. TF 3502 prints checkpoint messages.
The document provides explanations of these trace flags
Investigate SQL Server Memory Like Sherlock HolmesRichard Douglas
The document discusses optimizing memory usage in SQL Server. It covers how SQL Server uses memory, including the buffer pool and plan cache. It discusses different memory models and settings like max server memory. It provides views and queries to monitor memory usage and pressure, and describes techniques to intentionally create internal memory pressure to encourage plan cache churn.
The document provides an overview of performance tuning for Oracle databases. It discusses tuning goals such as accessing the least number of blocks and caching blocks in memory. It outlines the tuning process which includes tuning the design, application, memory, I/O, contention and operating system. Common performance issues for OLTP systems like I/O bottlenecks are also covered. Various tools for identifying performance problems are listed.
The document discusses several new features in Oracle Database 12c including:
- A new multi-tenant architecture using container databases and pluggable databases.
- Enhanced threaded execution that reduces the number of processes required.
- Ability to gather statistics online during direct-path loads instead of full table scans.
- Option to keep statistics on global temporary tables private to each session.
- Introduction of temporary undo segments to reduce undo in the undo tablespace.
- Ability to add invisible columns to tables.
- Support for multiple indexes on the same column.
- New information lifecycle management features like heat maps and data movement.
- Ability to log all DDL statements for troubleshooting.
- L
Oracle Database 12c includes over 500 new features. Some key new features include:
- Oracle Database 12c Express (EM Express) which replaces Database Control and has less features than Database Control but does not require Java or an app server.
- New online capabilities like online DDL operations with no DDL locking, online move of partitions with no impact to queries, and online statistics gathering for bulk loads.
- Adaptive SQL Plan Management which allows the optimizer to select a more optimal plan at execution time based on current statistics.
- Multitenant architecture which allows consolidation of multiple databases into one container database with pluggable databases.
The document describes the results of a proof of concept test comparing the performance of Oracle Database 12c In-Memory on Oracle SPARC M7 servers versus Intel servers. The test loaded and populated database tables with the TPC-H schema and benchmark and measured query throughput and response times under increasing user load. The SPARC M7 servers achieved significantly higher query throughput and lower response times compared to the Intel servers, demonstrating the benefits of the dedicated acceleration engines built into the SPARC M7 chip for in-memory processing.
The document provides an overview of performance tuning for Oracle databases. It discusses tuning goals such as accessing the least number of blocks and caching blocks in memory. It outlines the tuning process which includes tuning the design, application, memory, I/O, contention and operating system. Common performance issues for OLTP systems like I/O bottlenecks are also covered. Various tools for identifying performance problems are presented.
Healthcare Claim Reimbursement using Apache SparkDatabricks
The document discusses rewriting a claims reimbursement system using Spark. It describes how Spark provides better performance, scalability and cost savings compared to the previous Oracle-based system. Key points include using Spark for ETL to load data into a Delta Lake data lake, implementing the business logic in a reusable Java library, and seeing significant increases in processing volumes and speeds compared to the prior system. Challenges and tips for adoption are also provided.
Based on the popular blog series, join me in taking a deep dive and a behind the scenes look at how SQL Server 2016 “It Just Runs Faster”, focused on scalability and performance enhancements. This talk will discuss the improvements, not only for awareness, but expose design and internal change details. The beauty behind ‘It Just Runs Faster’ is your ability to just upgrade, in place, and take advantage without lengthy and costly application or infrastructure changes. If you are looking at why SQL Server 2016 makes sense for your business you won’t want to miss this session.
KSCOPE 2013: Exadata Consolidation Success StoryKristofferson A
This document summarizes an Exadata consolidation success story. It describes how three Exadata clusters were consolidated to host 60 databases total. Tools and methodology used included gathering utilization metrics, creating a provisioning plan, implementing the plan, and auditing. The document describes some "war stories" including resolving a slow HR time entry system through SQL profiling, addressing a memory exhaustion issue from an OBIEE report, and using I/O resource management to prioritize critical processes when storage cells became saturated.
This document summarizes the key points from a presentation on SQL Server 2016. It discusses in-memory and columnstore features, including performance gains from processing data in memory instead of on disk. New capabilities for real-time operational analytics are presented that allow analytics queries to run concurrently with OLTP workloads using the same data schema. Maintaining a columnstore index for analytics queries is suggested to improve performance.
GLOC 2014 NEOOUG - Oracle Database 12c New FeaturesBiju Thomas
This document discusses several new features in Oracle Database 12c for 11g DBAs. It covers improvements to administration and manageability such as online data file move, temporary undo, listener registration, and limiting PGA. New enhancements to tools and utilities are also presented, like Database Express 12c, enhanced patching, Data Pump NOLOGGING import, and improved features in RMAN. The document provides an overview of many updated features and directs readers to Oracle documentation for more details.
This document provides an overview of Oracle performance tuning fundamentals. It discusses key concepts like wait events, statistics, CPU utilization, and the importance of understanding the operating system, database, and business needs. It also introduces tools for monitoring performance like AWR, ASH, and dynamic views. The goal is to establish a foundational understanding of Oracle performance concepts and monitoring techniques.
En este diapositivas der Microsoft podemos ver qué aporta SQL 2014 en áreas como: Tablas optimizadas en memòria, Cambios en estimacion de la cardinalidad, Cifrado de los Backups, Mejoras en arquitectures, Always On, Cambios en Resource Governor, Data files en Azure.
- The document discusses various Oracle Database backup and flashback technologies including native backups to object storage, flashback query, flashback version query, flashback transaction query, DBMS_FLASHBACK package, flashback data archive, flashback table, flashback drop, and flashback database.
- It provides an overview of the requirements and examples for using each technology. It also includes a table comparing the availability of different flashback features across various Oracle Database editions.
- The document concludes with some best practices for backups and flashback as well as distinguishing flashback from database point-in-time recovery (DBPITR).
Similar to Oracle Exadata Performance: Latest Improvements and Less Known Features (20)
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/07/intels-approach-to-operationalizing-ai-in-the-manufacturing-sector-a-presentation-from-intel/
Tara Thimmanaik, AI Systems and Solutions Architect at Intel, presents the “Intel’s Approach to Operationalizing AI in the Manufacturing Sector,” tutorial at the May 2024 Embedded Vision Summit.
AI at the edge is powering a revolution in industrial IoT, from real-time processing and analytics that drive greater efficiency and learning to predictive maintenance. Intel is focused on developing tools and assets to help domain experts operationalize AI-based solutions in their fields of expertise.
In this talk, Thimmanaik explains how Intel’s software platforms simplify labor-intensive data upload, labeling, training, model optimization and retraining tasks. She shows how domain experts can quickly build vision models for a wide range of processes—detecting defective parts on a production line, reducing downtime on the factory floor, automating inventory management and other digitization and automation projects. And she introduces Intel-provided edge computing assets that empower faster localized insights and decisions, improving labor productivity through easy-to-use AI tools that democratize AI.
Hire a private investigator to get cell phone recordsHackersList
Learn what private investigators can legally do to obtain cell phone records and track phones, plus ethical considerations and alternatives for addressing privacy concerns.
Transcript: Details of description part II: Describing images in practice - T...BookNet Canada
This presentation explores the practical application of image description techniques. Familiar guidelines will be demonstrated in practice, and descriptions will be developed “live”! If you have learned a lot about the theory of image description techniques but want to feel more confident putting them into practice, this is the presentation for you. There will be useful, actionable information for everyone, whether you are working with authors, colleagues, alone, or leveraging AI as a collaborator.
Link to presentation recording and slides: https://bnctechforum.ca/sessions/details-of-description-part-ii-describing-images-in-practice/
Presented by BookNet Canada on June 25, 2024, with support from the Department of Canadian Heritage.
MYIR Product Brochure - A Global Provider of Embedded SOMs & SolutionsLinda Zhang
This brochure gives introduction of MYIR Electronics company and MYIR's products and services.
MYIR Electronics Limited (MYIR for short), established in 2011, is a global provider of embedded System-On-Modules (SOMs) and
comprehensive solutions based on various architectures such as ARM, FPGA, RISC-V, and AI. We cater to customers' needs for large-scale production, offering customized design, industry-specific application solutions, and one-stop OEM services.
MYIR, recognized as a national high-tech enterprise, is also listed among the "Specialized
and Special new" Enterprises in Shenzhen, China. Our core belief is that "Our success stems from our customers' success" and embraces the philosophy
of "Make Your Idea Real, then My Idea Realizing!"
In this follow-up session on knowledge and prompt engineering, we will explore structured prompting, chain of thought prompting, iterative prompting, prompt optimization, emotional language prompts, and the inclusion of user signals and industry-specific data to enhance LLM performance.
Join EIS Founder & CEO Seth Earley and special guest Nick Usborne, Copywriter, Trainer, and Speaker, as they delve into these methodologies to improve AI-driven knowledge processes for employees and customers alike.
Coordinate Systems in FME 101 - Webinar SlidesSafe Software
If you’ve ever had to analyze a map or GPS data, chances are you’ve encountered and even worked with coordinate systems. As historical data continually updates through GPS, understanding coordinate systems is increasingly crucial. However, not everyone knows why they exist or how to effectively use them for data-driven insights.
During this webinar, you’ll learn exactly what coordinate systems are and how you can use FME to maintain and transform your data’s coordinate systems in an easy-to-digest way, accurately representing the geographical space that it exists within. During this webinar, you will have the chance to:
- Enhance Your Understanding: Gain a clear overview of what coordinate systems are and their value
- Learn Practical Applications: Why we need datams and projections, plus units between coordinate systems
- Maximize with FME: Understand how FME handles coordinate systems, including a brief summary of the 3 main reprojectors
- Custom Coordinate Systems: Learn how to work with FME and coordinate systems beyond what is natively supported
- Look Ahead: Gain insights into where FME is headed with coordinate systems in the future
Don’t miss the opportunity to improve the value you receive from your coordinate system data, ultimately allowing you to streamline your data analysis and maximize your time. See you there!
Paradigm Shifts in User Modeling: A Journey from Historical Foundations to Em...Erasmo Purificato
Slide of the tutorial entitled "Paradigm Shifts in User Modeling: A Journey from Historical Foundations to Emerging Trends" held at UMAP'24: 32nd ACM Conference on User Modeling, Adaptation and Personalization (July 1, 2024 | Cagliari, Italy)
AI_dev Europe 2024 - From OpenAI to Opensource AIRaphaël Semeteys
Navigating Between Commercial Ownership and Collaborative Openness
This presentation explores the evolution of generative AI, highlighting the trajectories of various models such as GPT-4, and examining the dynamics between commercial interests and the ethics of open collaboration. We offer an in-depth analysis of the levels of openness of different language models, assessing various components and aspects, and exploring how the (de)centralization of computing power and technology could shape the future of AI research and development. Additionally, we explore concrete examples like LLaMA and its descendants, as well as other open and collaborative projects, which illustrate the diversity and creativity in the field, while navigating the complex waters of intellectual property and licensing.
Are you interested in dipping your toes in the cloud native observability waters, but as an engineer you are not sure where to get started with tracing problems through your microservices and application landscapes on Kubernetes? Then this is the session for you, where we take you on your first steps in an active open-source project that offers a buffet of languages, challenges, and opportunities for getting started with telemetry data.
The project is called openTelemetry, but before diving into the specifics, we’ll start with de-mystifying key concepts and terms such as observability, telemetry, instrumentation, cardinality, percentile to lay a foundation. After understanding the nuts and bolts of observability and distributed traces, we’ll explore the openTelemetry community; its Special Interest Groups (SIGs), repositories, and how to become not only an end-user, but possibly a contributor.We will wrap up with an overview of the components in this project, such as the Collector, the OpenTelemetry protocol (OTLP), its APIs, and its SDKs.
Attendees will leave with an understanding of key observability concepts, become grounded in distributed tracing terminology, be aware of the components of openTelemetry, and know how to take their first steps to an open-source contribution!
Key Takeaways: Open source, vendor neutral instrumentation is an exciting new reality as the industry standardizes on openTelemetry for observability. OpenTelemetry is on a mission to enable effective observability by making high-quality, portable telemetry ubiquitous. The world of observability and monitoring today has a steep learning curve and in order to achieve ubiquity, the project would benefit from growing our contributor community.
Details of description part II: Describing images in practice - Tech Forum 2024BookNet Canada
This presentation explores the practical application of image description techniques. Familiar guidelines will be demonstrated in practice, and descriptions will be developed “live”! If you have learned a lot about the theory of image description techniques but want to feel more confident putting them into practice, this is the presentation for you. There will be useful, actionable information for everyone, whether you are working with authors, colleagues, alone, or leveraging AI as a collaborator.
Link to presentation recording and transcript: https://bnctechforum.ca/sessions/details-of-description-part-ii-describing-images-in-practice/
Presented by BookNet Canada on June 25, 2024, with support from the Department of Canadian Heritage.
What Not to Document and Why_ (North Bay Python 2024)Margaret Fero
We’re hopefully all on board with writing documentation for our projects. However, especially with the rise of supply-chain attacks, there are some aspects of our projects that we really shouldn’t document, and should instead remediate as vulnerabilities. If we do document these aspects of a project, it may help someone compromise the project itself or our users. In this talk, you will learn why some aspects of documentation may help attackers more than users, how to recognize those aspects in your own projects, and what to do when you encounter such an issue.
These are slides as presented at North Bay Python 2024, with one minor modification to add the URL of a tweet screenshotted in the presentation.
Video traffic on the Internet is constantly growing; networked multimedia applications consume a predominant share of the available Internet bandwidth. A major technical breakthrough and enabler in multimedia systems research and of industrial networked multimedia services certainly was the HTTP Adaptive Streaming (HAS) technique. This resulted in the standardization of MPEG Dynamic Adaptive Streaming over HTTP (MPEG-DASH) which, together with HTTP Live Streaming (HLS), is widely used for multimedia delivery in today’s networks. Existing challenges in multimedia systems research deal with the trade-off between (i) the ever-increasing content complexity, (ii) various requirements with respect to time (most importantly, latency), and (iii) quality of experience (QoE). Optimizing towards one aspect usually negatively impacts at least one of the other two aspects if not both. This situation sets the stage for our research work in the ATHENA Christian Doppler (CD) Laboratory (Adaptive Streaming over HTTP and Emerging Networked Multimedia Services; https://athena.itec.aau.at/), jointly funded by public sources and industry. In this talk, we will present selected novel approaches and research results of the first year of the ATHENA CD Lab’s operation. We will highlight HAS-related research on (i) multimedia content provisioning (machine learning for video encoding); (ii) multimedia content delivery (support of edge processing and virtualized network functions for video networking); (iii) multimedia content consumption and end-to-end aspects (player-triggered segment retransmissions to improve video playout quality); and (iv) novel QoE investigations (adaptive point cloud streaming). We will also put the work into the context of international multimedia systems research.
Performance Budgets for the Real World by Tammy EvertsScyllaDB
Performance budgets have been around for more than ten years. Over those years, we’ve learned a lot about what works, what doesn’t, and what we need to improve. In this session, Tammy revisits old assumptions about performance budgets and offers some new best practices. Topics include:
• Understanding performance budgets vs. performance goals
• Aligning budgets with user experience
• Pros and cons of Core Web Vitals
• How to stay on top of your budgets to fight regressions
The Rise of Supernetwork Data Intensive ComputingLarry Smarr
Invited Remote Lecture to SC21
The International Conference for High Performance Computing, Networking, Storage, and Analysis
St. Louis, Missouri
November 18, 2021
An invited talk given by Mark Billinghurst on Research Directions for Cross Reality Interfaces. This was given on July 2nd 2024 as part of the 2024 Summer School on Cross Reality in Hagenberg, Austria (July 1st - 7th)