digital energy journal - issue 53
DESCRIPTION
Faster ways to find the rightvideoAcoustic fibre optics in wellsWhat you can do with seismic inthe cloudIs better subsurface knowledgethe key to improving thefinancial viability of offshoreoperations?TRANSCRIPT
-
February / March 2015 Issue 53
Faster ways to find the rightvideoAcoustic fibre optics in wellsWhat you can do with seismic inthe cloudIs better subsurface knowledgethe key to improving the financial viability of offshore operations?
Official publication of Finding Petroleum
-
Find out more and reserve your place at
www.d-e-j.comwww.d-e-j.com
Events 2014Non-seismic GeophysicsLondon, 19 Feb 2015
Finding Enough Oil & Gas in NW EuropeLondon, 12 Mar 2015
Transforming Sub-Surface InterpretationAberdeen, 17 Mar 2015
Doing more with Subsurface DataAberdeen, 18 Mar 2015
Transforming Sub-Surface interpretationLondon, 13 Apr 2015
Doing more with Subsurface DataStavanger, 05 May 2015
Doing more with Production DataStavanger, 06 May 2015
Finding Oil in Atlantic BasinsLondon, 27 May 2015
Events 2015
Find out more and reserve your place at
www.d-e-j.comAdmis
sion
from 20
www.findingpetroleum.com
-
Naturally, first reactions have been Give us a Tax
break! and How do we get Costs way down?
Yes, these are important because they finish up in the
Numerator of the crude economic equation that de-
scribes profitability.
But there is also a Denominator which is, or are, barrels
of oil or cubic feet of gas. How do we input more of
these into the equation?
I have two thoughts:
Beginning with a story. Many years ago I had a minor role in BPs takeover of Britoil
(previously of course BNOC, the UKgovernments national oil and gas company).
This takeover was underpinned by profound understanding of North Sea geology, of Yet-
to-Find volumes, of undeveloped discoveries, of upcoming development projects, of pro-
ducing fields.
In point of fact, BP probably understood Britoils acreage and fields better than Britoil
did itself.
This 1986 1988 work was almost entirely analogue paper composite logs, paper seis-
mic sections, hand-drawn maps, tracing paper, light-tables, occasional use of the digi-
tizing table (remember those?). It worked!
Today it is difficult, impossible actually, to see such a profound underpinning anywhere,
perhaps because lots of key individuals have moved on, perhaps because of lazy as-
sumptions that the North Seas best days are somehow behind it.
And yet the significant Johan Sverdup discovery in the NOCS, in a well-explored area,
was as I understand it, the result of deep geological knowledge and innovative thinking.
We know perfectly well how to do these things see for example this summary of how
work on Nova Scotia revitalised exploration there. www.qz.com/318755
Something similar, of similar scope and imagination, is needed for the North Sea and, ar-
guably, NW Europe as a whole. Somehow this has to be a multi-client study, driven and
delivered by oil and gas industry folk, not some academic or research exercise.
Secondly, we geoscientists have developed a lazy dependence on yet-another-towed-
streamer-3D-seismic-survey which we need to move beyond. There are all sorts of new
technologies out there, from seismic nodes, passive seismic, fibre optics, full tensor
gravimetry, electromagnetics, that can tell us much more about the sub-surface, bringing
better predictions, and higher volume successes.
In principle, digital technology should allow us to be more efficient and effective in inte-
grating diverse sources of existing data and new multi-measurements, thereby allowing
coherent interpretation from the scale of whole basins to individual producing reservoirs.
This deeper knowledge will indeed lead to wider opportunities.
Just one snag..
I havent yet seen a digital offering that can do this!
1February / March 2015 - digital energy journal
Deeper knowledge leadsto wider opportunitiesby David Bamford
Issue 53 February / March 2015
Subscriptions: 250 for personal subscription, 795 for corporate subscription.
E-mail: [email protected]
Future Energy Publishing Ltd39-41 North Road, London, N7 9DP, UKwww.d-e-j.com Tel +44 (0)208 150 5292Fax +44 (0)207 251 9179
Editor Karl [email protected] +44 208 150 5292
Conference ProducerPanas [email protected] +44 208 150 5295
Advertising, event sponsorship and Exhibitions ManagerRichard [email protected] +44 (0) 208 150 5291
ProductionWai [email protected]
Printed by Printo, spol. s r.o., 708 00 Ostrava-Poruba, CzechRepublic. www.printo.cz
Front cover art by Alex McKenzie, artist, Digital Energy Journal
If we do not act quickly, the UK Continental Shelf (UKCS) andNorwegian Continental Shelf (NOCS) will soon be on life support.
David Bamford is a director of Digital Energy Journal and Finding Petroleum, and a non-executive director of
Premier Oil
-
2 digital energy journal - February / March 2015
Leaders
Repsol and IBM - cognitive computing in E+PSpanish oil major Repsol is working together with IBM to build a $15m-$20m system to answer difficult questions about whichlicenses to bid for and how to optimise production, using experts together with big dataBy Santiago Quesada, Repsols director for exploration and production technology
Oil companies continue to make high-stake
decisions in the face of increasing uncer-
tainty and geological risk based on extremely
complex data sets.
Cognitive computing systems can help in ex-
ploration and production by helping individ-
uals to better interpret big data and then
make informed decisions based on that data.
As a result, companies can maximise access
to better exploration areas, increase the pro-
ductivity of maturing oil fields and their
value, enhance safety and mitigate environ-
mental risks.
Until recently, geoscientists have been tasked
with mostly manually reading and extracting
information from enormous amounts of data
including journal papers reports, seismic data
and models of reservoirs, wells and facilities.
Recognising the need for an intelligent solu-
tion, Repsol and IBM, leveraging years of
existing collaboration, recently teamed up to
develop cognitive technologies that can
analyse subsurface data in order to drive im-
provements in exploration and production.
Based at IBM's pioneering Cognitive Envi-
ronments Laboratory (CEL), the researchers
will work on two prototype applications
which are specifically designed to increase
Repsol's strategic decision-making in the op-
timisation of oil reservoir production and in
the acquisition of new exploration areas and
production fields, both onshore and offshore.
Repsol is making an initial investment of $15
million to $20 million to develop two appli-
cations with early results targeted for late
2015.
The team will work together in New York
and Madrid, with each company committing
six to 10 employees to develop the technol-
ogy.
People, devices and spaces
To best achieve this, the cognitive computing
technology infrastructure has been designed
to specifically interact with people across
various devices and physical spaces.
For example, the technology is able to
process questions asked by humans in natural
language and sifts through information to re-
spond with the most likely answers.
This, in turn, will enable individuals and
teams to make better decisions by overcom-
ing cognitive limitations posed by big data.
Scientists in the CEL will also experiment
with a combination of traditional and new in-
terfaces which are based upon gesture, robot-
ics and advanced visualisation and
navigation techniques.
Through these modalities, researchers can
leverage sophisticated models of human
characteristics, preferences and biases that
may be present in the decision-making
process.
The technology will also introduce new real-
time factors which should be considered such
as current news events around economic in-
stability, political unrest and natural disasters.
These tools are not intended to replace the
key stakeholders such as geologists, geo-
physicists, engineers, investment managers,
risk analysts and corporate strategists, but to
assist them with building more fluid concep-
tual and geological models, highlighting the
impact of the potential risks and uncertainty,
visualising trade-offs and exploring what-if
scenarios.
The new applications developed by Repsol
and IBM will improve the way oil companies
visualise and develop exploration and pro-
duction activities.
It is envisioned that companies from other
sectors will set up their own CELs to make
better informed decisions and, ultimately, in-
crease their companies bottom lines.
Santiago Quesada, Repsols director for explorationand production technology
Oceaneerings Video Vault - faster video retrievalOceaneering has developed the Video Vault cloud-based solution to make it easier to store, share and find the right piece ofsubsea and offshore videoSubsea engineering company Oceaneering
has launched a cloud-based live video
streaming and archiving solution called
Video Vault to make it easier to store, re-
trieve, view, share and analyze the video you
are looking for.
To simplify future video retrieval and analy-
sis, video files can be tagged geographically,
with comments or subtitles, or according to
the time it was recorded. This means you can
develop systems to quickly pull up
video relating to a certain location or
operational significance, and view
both real time and historical data.
The use of video in offshore opera-
tions, for both subsea and surface
monitoring, has grown massively in
recent years, and brought a lot of
challenges for finding the right video
image from hours of video files, says
-
Mark Stevens, Director of Communications-
GDS, Oceaneering.
Usually Oceaneering has about 100 simulta-
neous video feeds across the company.
The data files are getting larger all the time,
with more high definition and 3D cameras
and high-resolution video compression for-
mats.
Meanwhile there are many limitations to
physical media (such as DVDs, USB
drives). It is hard to manage and share files.
Common Operating Picture (COP)
Oceaneerings core aim is to help oil and gas
personnel create a Common Operating Pic-
ture. This is a term which originated in the
military, which means a view of all of your
operations on a (usually large screen) map,
where you can drill in to see what you need.
The Common Operating Picture serves as a
dashboard of all of your activity, showing
the locations of all your vessels and rigs.
When video is added to the COP, users can
click on any of numerous monitored assets
to see associated recorded video.
You can view a thumbnail of all your real
time video streams, displayed in the appro-
priate position of the map, refreshed every
30 seconds.
You can say show me all the video in the
Gulf of Mexico, then click on an individual
ROV in that area and say follow.
Some oil companies (including Statoil) have
their own 3D visualisation platform, and can
view this data in the COP, as well.
Benefits of video
Oil and gas companies are using video in
many different ways.
Subsea, companies use video for monitoring
ROV operations, observing rotary brush
cleaning operations, subsea X-ray imaging
(digital radiography) and general long term
asset monitoring.
The video is proving particularly useful in
subsea cleaning work. You can capture
videos before and after cleaning, to see how
effective the cleaning was, and then work
out a better cleaning schedule.
If the video is of a pipeline inspection, all of
the video can be geographically tagged, so
you can automatically bring up the video in-
spection image of a certain section of
pipeline.
A lot of companies are asking for video of
drilling data, particularly looking at the drill
floor, to record the drill pipe going through
the floor of the drill ship and back out again.
Viewing the data
In terms of viewing the data, about 30 per
cent of clients use mobile devices. That
continues to be the trend, Mr Stevens said.
It means staff can monitor offshore activities
from wherever they are, so less staff are re-
quired to be actually offshore.
To search for video on Video Vault, you start
with a Google Earth or ESRI map interface.
You can see all the live video streams with a
location on the map, or you can draw a box
around an area and see all the video you
have for that region. You can browse the en-
tire video library with advanced search, pre-
view and indexing, using thumbnails to
speed identification of interest areas, and
then play back selected video with the con-
venience of fast- and slow-forward, fast- and
slow-rewind, and pause capabilities.
Clients often ask to connect different files to
the video at appropriate points (such as pdf
documents and pictures), which you can do
with metadata content management.
You can provide usage data such as which
employees were watching the video and how
much bandwidth they consumed, as well as
total bandwidth utilization across the organi-
zation.
There are tools to deliver a short video file
to your desktop so if you want to watch a 5
minute section of a 4 hour video, you dont
need to download the whole thing.
Archival infrastructure
The Video Vault solution is made available
as a hosted cloud service for a tiered, daily-
rate or monthly price depending on applica-
tion, including maintenance. It is also
available as a bundled hardware solution for
on premise installation. This means you can
store the data on your local servers, or you
can host it on a standard commercial
archival service like Google. Clients never
have to see any physical storage media.
If you are concerned about storing your data
on standard commercial cloud services, you
could use Video Vault with Amazons Gov-
Cloud service, designed specifically for
more sensitive data.
Open standards / commercial systems are
used as much as possible (including integra-
tion with Google Earth and ESRI) to give
clients as much flexibility as possible. There
are application programming interfaces
(APIs) enabling you to connect Video Vault
with other software tools.
There are 10 software developers working
on the project, out of a total team of 55 de-
velopers at Oceaneering.
Geospatial Data Integration
Video Vaults ability to integrate geospatial
data means that the GPS co-ordinate of the
asset being monitored is automatically em-
bedded into the captured live video data
stream. For subsea monitoring, the GPS co-
ordinates of the topside vessel is used.
For some subsea projects, transponders have
been laid on the vessel floor, which enable
the gathering of location information from
the seabed.
Larger ROVs have an internal inertial map
(NAV) system, which estimate the ROVs
location based on how it is moving through
the water. It can calculate to within inches
where the ROV is, Stevens says.
In June 2014, Oceaneering acquired a com-
pany called PortVision, which has a vessel
tracking service and 5 year database of ves-
sel locations around the world derived from
automatic identification systems (AIS)data.
Oceaneering is now connecting the vessel
information with the video if you know
about the date and which asset the camera
was recording, you can put this together
with AIS data to work out the location.
3February / March 2015 - digital energy journal
Subsurface
Have a Common Operating Picture of your operations- drill down to get further information and live video
-
Interesting recent developments in land seismic recording technology include mini-vibes, low frequency seismic, and UAVsfor wireless seismic quality control, says Andy Bull, VP Product Development, Emerging Technologies at INOVA Geophysical.But we may be reaching the limit of the number of seismic channels we are able to record without significant advances incomputing technology.
Inova - mini-vibes, low fequency and UAVs
4 digital energy journal - February / March 2015
Subsurface
INOVA has developed a range of small Vi-
broseis trucks, known as mini-vibes, which
can create seismic energy in areas where
standard Vibroseis trucks cant go.
The usual way to create seismic energy on
land is by using large Vibroseis trucks which
provide 60-80,000 lbs of force.
Inova produces the UNIVIB and UNIVIB 2
Vibrator trucks, which provide up to 26,000
(UNIVIB) or 34,000 (UNIVIB 2) pounds of
peak force, but have the advantage of being
much easier to use in urban areas, or areas
with more difficult terrain, said Andy Bull,
VP Product Development, Emerging Tech-
nologies at INOVA Geophysical.
He was speaking at the Finding Petroleum
conference in London on November 26,
New E&P Technologies.
These little (mini vibe) machines are capa-
ble of really nice manoeuvring, low environ-
mental impact and tremendous broadband
energy, he said.
These machines can output super signals
over a bandwidth of 1-400 Hz and at signifi-
cant depth, he said.
In one example, data acquired in the South
Georgia Rift Basin
(USA), using 2 UNIVIB-
trucks, with data at 6 to
76 Hz. They have been
able to bring out a struc-
ture at 10,000 feet, he
said.
The market for mini-vibes
is partly driven by the in-
creased efforts oil compa-
nies have to make about
to limit the environmental
impact of exploration, he
said.
This includes thinking
about how much line cut-
ting (path clearing) they
need to do, how much dy-
namite they are going to
use, and what to do about
wildlife.
The environmental impact is further reduced
if you use cable-less technology and small vi-
broseis trucks, he said.
Because they have a shorter length and
width, less line cutting is required, he said.
They can also move up slopes of difficult ter-
rain.
INOVA Geophysical is a manufacturer of
land seismic technology, formed from assets
from ION Geophysical and BGP in March
2010. It produces a range of seismic record-
ing equipment and seismic source equipment.
Low frequency seismic
INOVA is seeing an increasing customer in-
terest in working with low frequency seismic.
We have been involved in several projects in
various locations where we starting with fre-
quencies as low as 1.5 Hz, and we think we
can go down further, he said.
MEMS (Microelectromechanical systems)
sensors are ideal for [recording] this low fre-
quency data, he said. The sensor systems are
designed to be small (107g), low power
(85mW) and tightly integrated with acquisi-
tion systems.
We're still learning a lot about the benefits
of these low frequencies.
Mr Bull showed an example of data surveyed
by Petroleum Development Oman (OMAN),
where during the survey, the company de-
cided to extend the bandwidth down to 1.5Data acquired in the South Georgia Rift Basin (USA) using 2 UNIVIB trucks was able to bring out a structure at 10,000 feet
Mini vibroseis machines - "capable of really nice manoeuvring, low environmentalimpact and tremendous broadband energy"
-
Hz, and you can see how the data improves.
You can see just the data generated from 1.5
6 Hz seismic and it is very clear.
The company has also done some downhole
testing with sensors recording at up to 7,500
feet, recording data as low as 0.5 Hz.
Downhole data was also recorded at high fre-
quencies, resulting in coherent 200 Hz en-
ergy visible at over 5,000 ft depth.
Seismic technology market
At the time of his talk in November 2014,
Inova was already seeing oil and gas compa-
nies make reductions in their exploration pro-
grams due to cost constraints due to the low
oil price, he said.
But the Middle East is still a healthy area.
There's a lot of investment still going on, a
lot of surveys planned, he said.
Annual spending on seismic technology has
been fairly flat for the past 10-15 years, with
a small uplift in the mid-2000s with a move
to larger surveys.
This creates difficulty for companies like
ourselves, he said. We're helping the mar-
ket acquire much more data, of better quality,
but the volume of spend is not increasing.
Many companies are starting to look seri-
ously at broadband seismic acquisition, as
they look for oil at greater depths, in more
complex geology, and with better resolution
images, he said. Or they want to do full
waveform inversion, using the two-way
wave equation to produce high resolution ve-
locity models.
Cabled and wireless
In terms of land seismic acquisition systems,
80 to 85 per cent of new seismic system sales
are still cabled, rather than cable-less.
The costs [of cabled systems] are reducing,
he said. The cable system is still substan-
tially less per channel than cable-less sys-
tem.
Cabled systems like INOVAs G3i HD need
to be able to run surveys at all kind of scales,
be integrated with other types of sensor (for
example with waterborne sensors for transi-
tion zone surveys), or used with cable-less,
and productivity is very important.
Keeping productivity high is a huge focus
for us, he said.
Meanwhile there is growing interest in cable-
less technology for environmental and safety
reasons, he said. It requires smaller crews.
The power consumption is reducing all the
time, so now devices are capable of 20-30
days operation in the field from one battery
charge.
With a cable-less system like INOVAs
Hawk, you can do powerful quality control
and noise monitoring simply, via wi-fi, he
said.
UAVs
Companies are now showing interest in de-
veloping unmanned aerial vehicles (UAVs)
which can fly around the spread of wireless
devices, download data over wi-fi and do
basic quality control on it, he said.
Purchasing a fixed wing UAV, which can
carry a 5 lb payload and fly for 2 hours, costs
about $150,000, he said.
But they might be able to quality control 9-
10 lines in a 2 hour period, compared to 1-2
lines using conventional methods.
By integrating UAVs with cable-less sys-
tems, powerful status QC and noise monitor-
ing can be achieved simply and even more
efficiently without the need for complex
radio infrastructure, he said.
But there are still regulatory obstacles to
using UAVs in many countries though,
he said.
Big data
There is a lot of talk about increasing the
number of channels (individually recorded
data streams) in a seismic survey, but people
have perhaps not taken into consideration
how they will communicate and store all the
data, he said.
With 250,000 channels, your data rate is 4 to
5 gigabits per second.
That's a tremendous amount of data coming
up the cable. The system has got to handle
this.
Typical tape drives can record 160 to 250
megabytes a second, and a solid state disk
can store 500 to 600 megabytes a second.
Disks in a RAID array could store 3 to 4
gigabytes per second.
But your data recording speed can be re-
duced by many factors including the proces-
sor, input-output, memory, cache, quality
control. You see that throughput onto your
recording device really comes down, he
said.
Were right on the edge of what's capable
and what's supportable in the field, he said.
Its a very big challenge - taking this acqui-
sition up to 500,000 and beyond that.
In terms of data storage, 500,000 channels
could generate up to 86 terabytes of data over
24 hours. Over 6 months, thats 12 petabytes
of data storage (one petabyte = 1000 ter-
abytes). A petabyte costs about $250,000
today. Its a massive cost, he said. We're
going to be really challenged to solve this.
Maybe we'll see a bit of plateau of what
people need in terms of channel count until
they understand how to handle this volume
of data.
5February / March 2015 - digital energy journal
Subsurface
View Andy Bulls talk on video and down-load slides at www.findingpetroleum.com/video/1095.aspx
Inova's G3i HD cabled seismic recording system - designed for high productivity
-
The best way to get the most value from seis-
mic data is to have a strategy for managing
the data at the beginning of the project, said
Jane Hodson, head of technical services at
DataCo, and formerly lead subsurface tech-
nologist with Centrica, speaking at the Digi-
tal Energy Journal Aberdeen conference on
November 27, Better ways to Manage Seis-
mic Data.
That will give a more value upfront than
managing it at the end.
Good data management is a key to opening
up that potential and we have excellent sup-
ply of subsurface data management expertise
in Aberdeen, Ms Hodson said. With this we
can be a global player in this field, we have
the domain knowledge, we have the people
and we have the technology.
As well as reducing the amount of budget
companies have available, the low oil price is
also likely to drive company towards acquisi-
tions, mergers and sales and even decommis-
sioning, she said, which all leads to big
demands on data.
With this current downturn in the market we
have to start looking at ways to work differ-
ently, work smarter, faster and work with the
data we already have, she said.
Seismic data at the right time
Having well managed seismic data starts
with how the data is delivered into your sys-
tem in the first place, she said.
A key question is who in the company first
receives the data.
Does new seismic data get delivered to your
data management team, to the IM team, or
does it go straight to the geophysicists? she
asked. Does it get catalogued?
Is the seismic data delivered with every-
thing attached, such as an acquisition report,
or a loading sheet? If it doesnt have a load-
ing sheet, how do you know what all the
data is?
Do you have the naming standards in place
so you can identify the type of seismic you
have?
Who quality controls the seismic
when it is uploaded? Is it the geo-
physicist? Is it the data manger?
How do you know that they have
loaded it correctly or have QC it?
Its imperative this is done, be-
cause if it is not then this hap-
pens.
I've had people coming to me and
saying, 'I cant see my seismic',
that they've loaded. I said, Youve
loaded it in Norway, but it's in the UK sector,
because you got the wrong UTM. So the
rule is get it loaded by someone who actually
knows how to do this not everyone has that
skill.
If you get the data catalogued, loaded, veri-
fied and ready to interpret then you have
done your job right.
Dry wells or worse
If you don't get the process right things start
to happen you wish had not.
The geophysicist does his/her interpretation,
creates all those lovely horizons and sur-
faces, and gives it to the geologist.
The geologist creates the geological model,
gives it to the reservoir engineer, and they all
decide where to drill the well. They drill the
well and then, oh, there's nothing there.
Back to the drawing board because you did
not ensure the data was loaded correctly in
the first place. Lost time, lost money and lost
opportunity.
Data verification and data integrity are key
things to be completed before any interpreta-
tion work is carried out then hopefully you
will have less failures.
Publishing and Archiving
Whether or not you find oil, the seismic in-
terpretation data needs to be captured and
published.
And when all this is done, and the entire
package is been tied up quite neatly, the last
bit is to archive or sell! Whether it is archiv-
ing in your own internal storage system or an
external one you need to make that decision.
Companies spends millions acquiring seis-
mic, and then they forget to do the critical
publishing of the final piece of work.
In the end it is essential you have a detailed
catalogue of your data, whether that is seis-
mic or well, or even documents.
Have your archive strategy in place, be-
cause you never know when you might need
to access it again.
As a final statement it means in the future
you can check your archives instead of
spending another five million pounds or
more on reshooting or purchasing seismic,
when you may have it already.
IT
So the geophysicist is on the workstation,
and wants to access the data, but [the net-
work is] so slow its driving them crazy.
You can do as much as you want to clean up
those data, but if the [network] you have got
is not up to the job, then whats the point?
Work with IT. Make sure that you have
everything in place like network connectiv-
ity. So when the geophysicists are actually
interpreting the data, they don't have to wait
10 minutes for each inline to display.
Geophysicists are a valuable commodity in
themselves so wouldnt you prefer they had
the tools to deliver the projects on time then
sit around waiting.
6 digital energy journal - February / March 2015
Subsurface
You can view Janes talk on video at www.d-e-j.com/video/1228.aspx
Getting more value from seismicIn the current financial environment, oil companies need to improve the way they get maximum value from their seismic. JaneHodson explained how to do this
Delegates at Digital Energy Journal's Nov 27 Aberdeen conference"Better ways to manage seismic data"
-
Using digital cables on the seabed, to record
seismic data and get a picture of how the oil-
field is changing as it is produced, is gener-
ally thought to enable an increase in oil and
gas recovery of 5 per cent over the lifetime
of the field, said Helge Brandsaeter, presi-
dent of OCTIO, speaking at the Digital En-
ergy Journal Stavanger conference on
December 10, People and Subsea Data.
But from an investment point of view, sys-
tems are not being installed because too
often it is seen as nice to have rather than a
necessity, he said.
Most oil and gas staff are busy meeting their
short term objectives, and dont have time
for longer term ones.
But the short term business case can be
much stronger if you see it as a way to mon-
itor the overburden (rock between the reser-
voir and seabed) and avoid costly
environmental fines, he said.
Octio is majority owned by Statoil Technol-
ogy Invest.
Overburden problems
The seismic recording can help spot prob-
lems leading to possible overburden leakage.
There have been a number
of well publicised cases in
Norway where a water in-
jection well put more pres-
sure into the rock than it
was able to handle, leading
to craters opening up in the
subsea, he said.
Fields in the Barents Sea
have shallow reservoirs, so
there is higher risk
You see faulting going
from the reservoir to sur-
face directly, he said. In
reservoirs like this, The
overburden is as important
as the reservoir."
In Brazil, the overburden
is young in geological terms which means
that if water is injected at too high pressure
you can open pre-existing faults to the
seabed, and water and oil seeping to the
seabed.
By monitoring the seabed with digital ca-
bles, you can see how the rock is fracturing
and if there is a chance the fracture will
reach the surface. Its not difficult to see
that seismic can map any such seepage, he
said.
You can see how the cracks develop subsur-
face. This will give the operational teams a
notice that something has to be done by the
injection, he said.
It is reasonably easy to set up a business
proposition which is reasonably sound, he
said.
Drill cuttings
The system is being used on the Oseberg
field (140km Northwest of Bergen), to mon-
itor the injection of drill cuttings and waste
water.
Drill cuttings are milled very finely at the
platform and injected into the reservoir.
This provides a much less expensive option
for managing drill cuttings than transport
back to land.
The average cost for a North Sea field for
transportation of cuttings and water waste is
close to $20m a year, he said. The alterna-
tive is to ensure safe injection, for $2m to
3m.
But Norway has zero tolerance for any
waste pollution. If you can't ensure safe in-
jection you have to transfer all the fluids on-
shore, he said.
Seabed infrastructure
Octio typically installs a mesh of seismic
sensors on the seabed, 50m apart, which
communicate with a hub and send the data
to surface.
One system has 172 sensors in a W shape
around the well on the seabed.
The sensors have 4 components, 3 compo-
nent accelerometers and hydrophones.
The sensors have an active mode for
recording seismic data in a survey, and a
passive mode for continually listening to
seismic data.
Everything is managed remotely, with data
7February / March 2015 - digital energy journal
Subsurface
OCTIO - business case for permanent reservoirseismicIt may be easier to justify spending on permanent reservoir seismic if it is used for overburden monitoring, or monitoring drillcuttings injection, rather than to monitor the reservoir, says Helge Brandsaeter of Octio
Octio produces a system with permanent digital cables on the seabed, whichcan make regular seismic surveys and monitor how the reservoir is changing.
Developing a better business case for permanent reser-voir seismic monitoring: Helge Brandsaeter, president,OCTIO
-
8 digital energy journal - February / March 2015
Subsurface
Using fibre optics to record seismic data in
wells has been done since 2010, but there
has been enormous progress made since
then, both with the technology and market
acceptance, said David Hill, chief technol-
ogy officer of UK company OptaSense,
speaking at the Finding Petroleum London
conference on Nov 26, New E&P Tech-
nologies.
Seismic recording with optical fibre using a
technique called Distributed Acoustic Sens-
ing (DAS) in wells is not as sensitive as con-
ventional geophones, but the target is to
reach that level, he said.
Some people say the oil and gas industry
takes 30 years to adopt a new technology,
but the technology has already been widely
implemented in 5 years, he said.
You dont need to drill a special observation
well, you just install a fibre-optic cable in an
existing producer or injector well. If there is
already a fibre-optic cable in the well (per-
haps installed to monitor temperature), you
can use that.
The DAS technology has been used on one
offshore well so far, operated by Shell in the
Gulf of Mexico. Shell was conducting a
large seismic survey in the region and
wanted to see if it could get any value from a
fibre-optic cable which had been installed in
one of the wells a few years ago for a differ-
ent purpose, he said.
We were asked to attach our box to it, to
see if we could get any usable seismic.
The recorded data was so good, that one of
our employees had to spend seven weeks on
that platform, including over Christmas,
recording 50,000 shots.
1400 channels were recorded simultane-
ously, from two 6km fibres.
In another project, in a mature field in
Oman, the customer wanted to monitor
where injected steam was going, so they
could plan an infill well strategy. To do this
fibre was installed in eight wells, a task
which took about a month, and then a 3D
DAS-VSP (Vertical Seismic Profile) was ac-
quired simultaneously on each well
The data was correlated and stacked as it
was being recorded, enabling quality control
do be done immediately.
After seismic shooting, the data is being
stitched together to get a field wide view, he
said.
The same system can be used for time lapse
seismic, comparing a seismic survey today
with the seismic survey at some time in the
past.
For this to work, the fibre probably needs to
be fixed firmly in the well (cemented or at-
tached to the casing or production string),
not dangled on a wireline.
The technology could also be used to moni-
tor CO2 storage wells, he said.
The repeatability of the recorded data helps
people gain confidence in it, he said.
The fibre can also be used for microseismic,
recording natural seismic energy, which can
also be used to understand the subsurface.
Recording passive seismic data in an oil well
is not a new idea. It is conventionally done,
by drilling an observation well and inserting
geophones in it, or installing the geophone in
an existing well. The work is risky and ex-
pensive, he said.
In one example, fibre was installed on a
highly deviated well, which meant that by
triangulation you could work out where the
source was.
Optasense currently has 180 employees and
offices around the UK, Houston, Calgary,
Dubai and Australia. It recently acquired two
California based companies; RIO (Redfern
Integrated Optics), which produces the spe-
cial laser which sends the light through the
fibre, and SR2020, a specialist in high defi-
nition borehole seismic imaging and inter-
pretation.
How it works
The technology works by firing a pulse of
light (laser) into the fibre.
The glass fibre is the purest material man
has ever made, but there is enough inhomo-
geneity in the molecular structure to cause a
small amount of light backscatter, he said.
The backscatter appears to be random, but it
stays relatively constant if the fibre is not
disturbed.
Using acoustic fibre optics in wellsUK company Optasense reports that enormous progress is being made with using fibre optics in wells for recording seismicdata, monitoring, fracking and flow, among other applications.
transmitted back to shore.
The installation of the system basically
comes down to the cost of leasing vessels,
and installation cost is a third of the lifetime
operating costs.
The data communications infrastructure can
also be used for any other subsea equipment.
If you build such an area wide infrastruc-
ture we can use it for all types of communi-
cations and types of sensors, he says. You
can drop down a sensor and communicate to
surface.
We have made basically an ethernet on the
seafloor, you can interface any system to
us, he said.
Octio is working with a number of standardi-
sation committees including SWIG (Subsea
Wireless Group) and SIIS (Subsea Instru-
mentation Interface Standardisation).
Watch Helges talk on video at www.d-e-j.com/video/1502.aspx
Making progresswith recordingseismic in wellswith fibre optic David Hill, chieftechnology officer,Optasense
-
9February / March 2015 - digital energy journal
Subsurface
CGG - time for seismic on the cloud?For economic reasons if nothing else, the oil and gas industry is likely to start moving its seismic data to the cloud, saysCGGs Henri Blondelle
But if there is a tiny strain on the fibre,
which can include a strain caused by a noise,
the backscatter pattern changes slightly.
With a calculation involving the speed of
light, you can calculate which part of the
fibre that event happened.
The processing work will typically divide a
length of fibre into a number of sections and
work out the sounds which could be heard at
each section of the cable (so it is equivalent
to one seismic channel). For example a 5km
fibre can be split into 500 x 10m sections.
The fibre can be installed by attaching it to
the production string, or cemented behind
the casing. It can also be run into the well on
a wireline or slickline. It has been used on a
well 7km long.
The fibre-optic cables are very robust.
Fibre is actually stronger than steel, he
said. These cables have been developed
over 20 years.
The fibre-optic cable can handle tempera-
tures of up to 300 degrees C.
The fibre response is quite directional, and
not so sensitive cross axis so it cannot de-
termine which direction a sound is coming
from. One possible solution is to wind the
fibre around the well, he said, or try to de-
velop a perpendicularly sensitive cable.
Fracking and flow
The DAS acoustic data can be used to moni-
tor what is happening in the well, including
fracking operations, monitoring flow, water
loading and valve operations.
For example you can see a rising water level
in the wells. You can also record the flow
rate of oil and see the trends.
Sometimes wells gradually fill with water,
and then the oil suddenly pushes past it,
sending the water back into the reservoir.
You can monitor what is happening across
inflow control valves (ICVs) and at the gas
lift values (GLVs) in order to make sure they
are working properly.
You can also monitor pressure. In the future
we hope this technology will replace gauges
downhole, he said.
The technology has been used to monitor
electrical submersible pumps (ESPs),
analysing the noise to see rotational speeds,
and spot for any cavitation effects in the
motor.
In fracking, you can check the perforation
charges are firing properly, and then monitor
the flow through the perforations. You can
also listen to what is happening with a frac
job on a neighbouring well.
Sometimes frackers push proppant into an
adjacent well, potentially blocking it, he
said.
It has been used to make sure plugs are set
properly, because of a particular sound sig-
nature which is heard when a plug is set.
Sometimes, the system detects wells having
sudden leaks, releasing fluid at high pressure
and then closing. This is something a stan-
dard sensor would probably not capture, he
said.
The system cant get do multiphase flow
measurements but there are strong indica-
tions it might be possible, he said.
The technology has nowhere near fulfilled
its full capabiilty yet.
Data
The company is trying to come up with a
standard way to define the data, so the data
can be transferred between systems.
The data files can be enormous, with one
well generating a terabyte a day without
any problem, he says. That's only going to
get orders of magnitude worse as technology
progresses.
To keep data files manageable, it is essential
to process the data at source, so you are only
transferring the much smaller processed data
files, he said.
In 2015, version 4 of the OptaSense DAS
system will be made available, with an extra
6dB signal to noise improvement and better
spatial resolution, he said.
Other applications
Optasenses biggest business application for
the technology so far is for pipeline monitor-
ing. It is currently installed on 12,000 km of
pipeline, to detect potentially damaging ac-
tivity along the pipeline and detect leaks.
The technology is also used to monitor and
installed around factories to monitor for peo-
ple climbing over or cutting fences.
It is being used for condition monitoring, to
monitor condition of risers.
View a video of Davids talk at www.find-ingpetroleum.com/video/1060.aspx
Many companies in the retail and banking
sector moved their data to the cloud in the pe-
riod 2008-2009, driven by the need to reduce
costs after the 2008 crisis, said Henri
Blondelle, VP Global Business Development
CGG Data Management Services.
Perhaps now oil companies will move seismic
data to the cloud for the same reason.
He was speaking at the Digital Energy Jour-
nal conference in Aberdeen on November 27,
Better Ways to Manage Seismic Data.
The benefits of cloud data are well publicised,
but perhaps none of them motivate activity as
much as short term economic gain.
The financial case of storing data on cloud
servers, rather than on your own, is
quite strong.
If you want to store data in-house, your in-
vestment in buying storage equipment is ini-
tially quite high, followed by a small
additional cost each year for new data stor-
age. So your purchasing costs are uneven.
You might choose to buy new equipment
-
again after 3-5 years.
Using a cloud or provider you can keep the
cost the same every year, and the cost will
also decrease as the cost of storage equipment
decreases. You only pay for the storage you
need, and can increase or decrease it on de-
mand.
I have seen a lot of pilots, a lot of proof of
concepts, which for me prove that the indus-
try will be ready in the next few months to
absorb this challenge [of putting seismic on
the cloud], he said.
Once the data has been moved onto the cloud,
the industry can gain many further benefits,
already used by other industries, such as the
ability to run Hadoop based interpretation on
it, he said.
CGG works together with Microsoft Azure as
a cloud service provider, and proposes its cus-
tomers a package with CGG Data Manage-
ment services running over Microsoft Azure.
Security and control
Many oil and gas companies have chosen not
to move data to the cloud due to security con-
cerns, an argument made stronger by many
recent cloud hacking stories.
To try to mitigate these concerns, cloud serv-
ice provider Microsoft Azure recently con-
ducted a survey of its existing customers to
ask them what they thought about security
after having moved their data to the cloud.
All of them considered that the security,
after moving to the cloud, is better than be-
fore, Mr Blondelle said.
A majority (but not all) users consider that
they have more control on the data on the
cloud than with their previous (internally
hosted) solution, he said.
There are many ways cloud service providers
can improve security, such as storing data en-
crypted, ensure the security of the data centre,
and find ways to transfer the data securely, he
said.
Some companies might have legal require-
ments to store data in a certain country, or
might choose not to put some data on the
cloud because of slow internet connections.
Some of these problems can be resolved by
having a partial cloud solution, he said.
Of course the cloud system can make it much
easier to provide other parties with access to
your data as required, including governments
and joint venture partners.
Easier to work with
One often overlooked advantage of putting
seismic data in the cloud is that it is easier to
work with in many ways, he said.
For example, you can create automatic tools
to monitor the quality of your data, and run
business rules. One oil company has 6,000
business rules.
You can display data from many different
servers together, for example if you have sep-
arate databases for well logs and seismic data
files.
You can keep track of all the data processing
tasks which have been done, which is very
helpful if you have to discover an error some-
where in the process you want to roll back,
and have had several interpreters working on
data simultaneously.
You can use sophisticated cloud based analyt-
ics tools which can interpret text.
You can use analytics tools such as Tibcos
Spotfire or Tableau Software. These tools
make it easy to share the results of the analyt-
ics unlike if you do the analytics running
over Petrel, which means that someone needs
to login to a Petrel workstation to view them.
Many interpreters prefer to use something
that looks like Excel - like Spotfire or
Tableau, he says.
Teradata and Horton Works provide a range
of statistical tools you can use.
You can store your seismic data in a Hadoop
based storage solution, which gives you the
ability to do some Hadoop analytics without
downloading the data, including all the seis-
mic processing and interpretation.
In one example, Statoil used Hadoop cloud
storage to make correlations between changes
in seismic data and pore pressure calcula-
tions.
Well logs and drilling data
CGG ran a pilot project in mid-2014 together
with Teradata mixing well logs and drilling
data covering the whole of the North Sea.
It worked with a very difficult drilling data
set, mainly pdfs of well reports, with mud
density data on them. Its very difficult to
extract this information from a pdf, he said.
The project team also tried to automatically
extract data from well logs and well comple-
tion reports.
After extracting the data, efforts were made to
try to find correlations.
Some of the correlations were as expected.
The drillers know already the formations
which are difficult to drill, he said.
But there was also some detailed information
which could be very useful. In this part of
the North Sea, for this particular formation, if
you use this particular mud density, you have
80 per cent of chance to have a bad hole sec-
tion, he said.
There is a possibility of bringing this quanti-
tative information to the people who have to
make a drilling program.
Indexing
CGG has also looked at ways to make seis-
mic data easier to catalogue and index auto-
matically.
Indexing data the classical way generally
involves opening it, reading it, and entering
the relevant information in a database, which
can be time consuming, if you have millions
of files to index. As a result the task is often
not done.
So perhaps if the documents could be
analysed and indexed automatically, oil com-
panies would index documents which other-
wise would not be indexed at all.
These tools are already used by US law firms
to automatically classify reports on
cases, he said.
Subsurface
10 digital energy journal - February / March 2015
View Henris talk on video at www.d-e-j.com/video/1223.aspx
One seismic data hasbeen moved to thecloud, you can gainmany extra benefits,such as being able todo Hadoop processingand automatic data in-dexing - HenriBlondelle, global busi-ness developmentmanager, CGG DataManagement Services
-
The classic internet search engine, digital li-
brary and enterprise search have traditionally
focused on precision and ranking.
The rationale is that as long as the specific
web page or document you were seeking is
on that first page, it does not matter how
many results are returned.
This approach has been incredibly successful,
leading to Internet search engines like Google
attracting a crowd nearing one billion users a
week, of which 94 per cent never click past
the first page of search results.
But increasingly with Internet search, smart
algorithms recommend or suggest related in-
formation, trying to predict what we need or
may find interesting.
In addition, social networks undoubtedly aid
discovery. However, some researchers feel
the overuse of historical usage and activity
data within algorithms to make suggestions
may place us in a filter bubble constraining
some potential serendipitous encounters.
Enterprise search
In an enterprise environment, significant frus-
tration still exists where the success seen on
the Internet seems harder to replicate inside
an enterprise.
Factors for unsatisfactory retrieval include in-
vestment levels, organizational culture, the
nature of workplace tasks, information gover-
nance and interventions, small crowds, infor-
mation structure and permissions along with
information behaviours of staff and manage-
ment.
Exploratory search
Exploratory search is where the question is
not fully formed in the mind of the searcher.
This is different to known item (or lookup)
search.
It is possible the actual need may in part be
stimulated by the search engine itself, with
the search engine acting like a creative mem-
ber of the team making suggestions from ini-
tial inputs.
Faceted search
Faceted search shows a breakdown of what
exists in the search results by various cate-
gories with counts, normally shown on the
left hand side of the screen inviting further
human interaction to browse and filter results.
These may be potentially useful options when
you consider most enterprise searchers enter
two words or less, searching increasingly
larger haystacks of information, so most
searches deliver hundreds or thousands of re-
sults.
But, it is difficult to represent the richness of
a 50 page report with 6 topics. Furthermore,
the same information item will always be rep-
resented by those same 6 topics, regardless of
what search terms are used and where rele-
vant matches are found inside the document.
One method to provide contextual based
topic filters is word co-occurrence - using
words that appear in proximity to the search
terms found in documents.
Need to be surprising
Recent research by Robert Gordon University
published in the Journal of Information Sci-
ence identified certain information needs with
respect to faceted search refiners.
Research was conducted using word co-oc-
currence stimuli generated from data pro-
vided by the Society of Petroleum Engineers,
Geological Society of London and the Ameri-
can Geological institute. The stimuli was
used to gather survey data from 54 petroleum
engineers from over thirty oil and gas indus-
try organizations.
A need was identified for the surprising as a
search filter.
The research found the most statistically fre-
quent associations (to search terms) were
often too vague and no promise of telling
me anything I didnt already know, rele-
vant but not interesting and contained few
surprises.
However, algorithms such as mutual informa-
tion measure appeared to generate more in-
triguing associations useful for deep dives,
might learn something and high on inter-
estingness quotient, you cant say where
these results may lead you.
Algorithms for surprising
Further research presented at the International
Conference on Knowledge Management used
discriminatory word co-occurrence tech-
niques surfacing potentially surprising asso-
ciations to search terms.
Initial results were promising. In an observa-
tional study of 53 geoscientists in two oil and
gas organizations, 41 per cent felt current
search interfaces used by their organization
facilitated serendipity to a moderate/large ex-
tent, increasing to 73% with the introduction
of certain algorithmically generated filters.
As put by one participant Its like open up
the box for me and Ill pick what does not fit
with my brain, like one of those games.
Surprising and serendipitous encounters oc-
curred giving rise to learning experiences, It
is clear I underestimated the importance of
carbonates in this is immediately important
for the research I am undertaking now.
Surprising associations can be unusual words
or quite common words but appearing in an
unusual or discriminatory context.
For example, What is interesting is that
Halite is there for the Permian, but techni-
cally it could occur for Tertiary, Jurassic,
(others), what is surprising is that it has not.
11February / March 2015 - digital energy journal
Subsurface
Using surprise in subsurface knowledge researchUntil now, enterprise search systems have focused on precision helping you find exactly what you want. But our internetsearch engines are moving more and more towards other ways of guessing what you might want. And surprising you withwhat they find. Should enterprise search move in the same way? By Paul Cleverley and Simon Burnett, Robert GordonUniversity, UK
Subsurface searchneeds to guess whatpeople might be look-ing for - and surprisethem with somethingelse - Paul Cleverley,researcher at RobertGordon University
-
12 digital energy journal - February / March 2015
Subsurface
Hadoop is a new data storage technology
which may lead to a big change in how the oil
and gas industry manages big data.
High-volume data, whether structured or un-
structured, can be placed into an enterprise
data hub (EDH) on Hadoop and prepared for
further analysis without the time-intensive
steps associated with the traditional relational
database model.
With a Hadoop cluster, both structured and
unstructured data can be managed for use
over long time horizons, in original fidelity,
and integrated with existing infrastructure and
tools.
With Hadoop, organizations have a new way
to think about data, transforming it from a
cost to an asset.
Geophysicists are already evaluating Hadoop
as a technology to support processing seismic
data from a multidimensional perspective,
complementing the traditional processing se-
quence.
New technologies that reduce the latency of
data for analysis, such as cloud applications
and Hadoop, are significant game changers
for the oil and gas business, said Moray
Laing, SAS Executive lead for Oil and Gas,
formerly at Baker Hughes.
As a result, this industry is on the precipice
of major change in their IT architectures.
Big data trends
Recently, The Data Warehousing Institute
(TDWI) released "Managing Big Data," a re-
port that explored trends in big data manage-
ment.
The report presents the findings from a sur-
vey of more than 400 practitioners about their
big data efforts.
Respondents were asked which database
management systems (DBMS) were in use
for big data management efforts.
While traditional relational DBMS systems
were at the top (38 percent), Hadoop was tied
for second place (33 percent).
In addition, it was clearly evident that the
phrase big data was synonymous with
Hadoop, in the minds of those surveyed. A
meteoric rise for a technology that only be-
came commercially viable in the last five
years.
For many organizations, establishing an en-
terprise data hub using Hadoop will be a cost-
effective solution for data capture of all data,
structured and unstructured, in a secure, man-
aged environment. When paired with addi-
tional technology applications to ensure data
quality, and to visualize and analyze the data
effectively, Hadoop is ready for prime-time.
Software companies, like Cloudera and SAS
are working together to provide processes and
technologies that accelerate data-driven in-
sights.
Dave Cotten, whose team at Cloudera sup-
ports many US oil and gas companies, says
that Cloudera's oil and gas clients are realiz-
ing multiple revenue generating and cost sav-
ings opportunities.
From real-time field operations feedback im-
proving reservoir yields, to full-fidelity elec-
tronic well record management, to mining
internal and public data to determine optimal
well spacing, customers are obtaining deeper
insights at lower costs provided by Hadoop in
an enterprise data hub.
In addition, our customers typically improve
preventative maintenance, greatly reducing
costly downtime.
Analytics on Hadoop
SAS and Cloudera recently announced tech-
nologies that move the analytic functions di-
rectly within a Hadoop cluster.
Deploying models directly in Hadoop reduces
data movement and replication, saving time
and resources - while strengthening data gov-
ernance.
With all your data in one place, simple tabular
data can mix with more complex and multi-
structured data to provide business insights
never before possible.
Organizations can run a variety of enterprise
workloads, from batch processing to ad-
vanced analytics, in a secure, managed, gov-
erned environment.
Early adoption of analytics on Hadoop has
been popping up in seismology, asset opti-
mization, commodity pricing strategies, and
production optimization.
Because SAS data visualization on Hadoop
allows companies to interactively explore bil-
lions of rows of data in seconds.
One common use case is for data validation,
finding the outliers and flagging them for fur-
ther explanation.
Because you can look at both structured and
unstructured data, like Twitter feeds or web
traffic, in one place over time, there are appli-
cations for cybersecurity as well.
Hadoop, a new data storage technology, promises to make high volume data much easier to manage in the oil and gasindustry. Alyssa Farrell of SAS explains how it works
Hadoop for oil and gas
This may be detached from any initial spe-
cific intent, the surprising nature of the asso-
ciation enticing the searcher to drill down
further which may lead to a serendipitous en-
counter.
Enhancing creativity
What is deemed surprising or intriguing by
one person, may not be by another as sug-
gested filter terms are compared with their
own cognitive map, like a game of spot the
difference.
The challenge with text co-occurrence is to
decide what to present to the user, minimiz-
ing distraction but offering potential sur-
prises, combining with traditional controlled
vocabulary (taxonomy) metadata approaches.
Companies that adopt such practices, may ex-
perience more happy accidents in the user
interface than those which do not.
Paul Cleverley and Simon Burnett areresearchers in the department of Infor-mation Management at the AberdeenBusiness School at Robert Gordon Uni-versity in Aberdeen, UK.
-
At the recent Society of Exploration Geo-
physicists (SEG) event in Denver in October
2014, a company called Agile Geoscience ran
a hackathon with 30 people in a room, asking
them to write a supercool geoscience appli-
cation.
David Holmes, one of the judges, selected as
the winner a crowdsourcing hot or not tool
for other peoples seismic interpretation.
You log on to an online tool with your Google
account, interpret some seismic data, and then
rate other peoples choices.
This is the future, I'm convinced, Mr
Holmes said, speaking at the Digital Energy
Journal Aberdeen conference on November
27, Doing more with Seismic Data.
There are many more exciting things the in-
dustry could do, if the seismic data systems
were on the cloud.
One US company put all of its seismic data
onto disk, and then hired 6 students from the
Colorado School of Mines and gave them ac-
cess to the entire seismic library, telling them
go and find some stuff.
Mr Holmes said it is a mystery to me why
we keep standalone workstation going as long
as we have, he said.
Geophysicists still work on personal worksta-
tions, where they spend 20 minutes loading
up all their data every morning. If they could
work directly on a cloud system it would be a
lot faster.
Managing old seismic
Yet still most oil and gas companies store
their seismic data on tape, and have very little
idea what they have, if they are storing multi-
ple copies of the same data, or do not have
what they thought they had.
Companies should care more about this
stuff. They have spent millions on acquiring it
and the cost of managing it is an unmeasur-
able fraction of that, he said.
Some companies do have a regulatory obli-
gation to keep their data in perpetuity, and it
doesnt mean a rusty 9-track no-one can
read.
A [typical] oil company is working with five
seismic data storage companies, each with
different cataloguing systems, all incompati-
ble, three million media items, including two
million 9-track tapes.
There are a finite number of read heads for
9-track tapes, he said. They are not being
manufactured any more.
Some companies are paying $10m a year in
license fees for data they're not using but can't
prove that they're not using, he said.
The trouble is, managing data is hard work,
and it always easier not to do it. Companies
ask, shall we spend lots of time and money
doing something hard with intangible busi-
ness value, or do nothing, he said.
Some physical data storage companies are
taking advantage of oil companies willingness
to take the cheap and secure option over the
short term. They offer a service where they
store your data free of charge, but charge you
big fees when you want to retrieve it, he said.
Some cloud data services are trying to get
away with the same business model. The
cost of retrieving the data can be gigantic.
The problem is that usage rates of seismic
data is typically very low, with only small
amounts of data retrieval over a time frame
measured in decades, he said.
But slowly, attitudes are changing, as compa-
nies realise the risk of not properly maintain-
ing data assets, he said, and cloud solutions
offer a cheaper alternative.
Move away from tape
Mr Holmes recommendation is to move away
from tape.
You cant just copy seismic data from tape to
disk, because the data will get corrupted. It
needs to be transferred to a different format.
Together with the data, you need to keep a
scanned copy of the tape label, logs of the re-
mastering process (gathering data from tape).
You need everything in your possession that
will allow you to recreate that nasty crumbly
9 track tape, he said.
Data management processes
In the future there will also be much stricter
data management processes, he said.
Many oil and gas companies already say that
geoscientists may not load up data them-
selves, they must give it to a data manager, to
load it and validate it, he said.
There are software tools to make this process
easier, for example where new data is loaded
to a folder, then a data manager receives an e-
mail alert. The data manager can then check
the data formatting and co-ordinates, make
any necessary transformations and check the
headers.
Working with big data
A side-effect of the growth of big data sys-
tems is that many companies now have multi-
ple systems for storing data, including their
normal archiving systems, high performance
computing (HPC) enviroments, Hadoop envi-
ronments.
They might have the same data file in all of
these systems. If they back up the data in each
environment multiple times, they can end up
with many copies of the data. One company
worked out they would have 17 copies of all
of their data, if everything had gone well, he
said.
As data volumes get bigger, keeping 17
copies of everything will get very expensive.
If we have any chance of surviving the next
few years, its going to be crucial that we
have a single instance of our data, he said.
Or companies will make a fortune selling
you vast amounts of storage you don't need.
A new term has been invented, next genera-
tion data fabric, which describes the enter-
13February / March 2015 - digital energy journal
Subsurface
EMC and the future of seismic dataDavid Holmes, chief industry executive with EMCs Global Oil & Gas Program, explained what the future of seismic datamanagement will look like
Could future seismicinterpretation couldbe done by crowd-sourcing, askedDavid Holmes, chiefindustry executive,EMC's Global Oiland Gas Program
-
The agreement is for LMKR's "GeoGraphix"
subsurface interpretation system to integrate
with petroWEB's "Enterprise DB" exploration
and production data management system,
built on the PPDM model.
Enterprise DB can serve as a corporate well
master, well log repository and well file man-
agement system.
By putting Enterprise DB together with Geo-
Graphix, you have a single system for manag-
ing large volumes of subsurface data, LMKR
says.
LMKR has also formed a technology partner-
ship with LUMINA Geophysical, a company
based in Houston which provides special
tools for quantitative interpretation of the sub-
surface, based on a mathematical method
called spectral decomposition. This allows
more geological information to be extracted
from geophysical data.
As a result of this partnership, LMKR will re-
lease a new software tool called Predict 3D in
early 2015. This is a multi-attribute inversion
solution based on spectral decomposition.
The multi-attribute inversion process predicts
rock properties from well logs away from and
in between the wells that have been used for
correlation. This helps understand the nature
of the reservoir away from the wells.
LMKR has also made improvements to the
user experience with its GeoGraphix 2014.2
subsurface software, improving seismic inter-
pretation speed and capability, offering larger
memory access and up-front well loading op-
timization.
There is improved integration with depth geo-
models. This means that users can easily
depth convert faults and incorporate into a
sealed earth model as well as combining seis-
mic interpretation data and well picks into the
geomodel.
Enhancements to velocity modelling and
depth conversion make it easier to depth con-
vert seismic with velocity models generated
from defined horizon-formation top relation-
ships.
LMKR has developed a 3D volume attribute
generation tool, which can manage both trace
based attributes (such as sample, window and
Hilbert), and frequency based attributes. All
of the frequency-based attributes use a
patented spectral decomposition technique for
higher resolution results.
LMKR GeoGraphix 2014 also delivers tighter
integration between geological and geophysi-
cal interpretations.
14 digital energy journal - February / March 2014
Subsurface
Developments at LMKRSubsurface data and modelling company LMKR reports that it has formed a partnership with petroWEB, an oil and gas dataand information management company based in Colorado, Canada and Houston.
prise architecture for storing and managing
information, he said.
Companies will also use object storage
which means that the analytics tools can un-
derstand the different data storage systems
you are using.
The idea of master data management will be
redundant, because companies will be able to
search all of their data at once.
Geophysicists will be able to ask complex
queries, like show me all the files I have
navigation for, which I dont know about. Or
in more specific terms, show me all of the
navigation files which have a survey name
which isn't in my survey master.
You can run that simple query against
your entire landscape, he said.
"Optique", a 4 year, $17.5m oil and gas big
data research project co-ordinated by a Uni-
versity of Oslo professor, is now in the mid-
way stage. The project has EUR 9.7m
funding from the European Union and is sup-
ported by DNV GL, Siemens, Statoil, and a
German cloud data company called "fluid
Operations".
Optique aims to develop a software platform
to help oil and gas users work with large and
complex data sets, using "semantic technolo-
gies", which includes data meaning as part of
the data model.
The project was initiated in 2010 by Univer-
sity of Oslo (UiO) professor Arild Waaler, and
launched in 2012. Total funding is Eur 13.8m
(USD 17.5m) including Eur 9.7m from the
EU.
Professor Waaler believes that this system
will be different to other big data solutions, in
that it will focus on understanding the com-
plexity (including the variety) of the data,
where most other big data solutions just focus
on working with large data volumes.
"Optique .. addresses trustworthiness by
showing where data came from and how it
has changed, providing transparency for the
end user," he said.
At the moment, geologists and engineers need
to involve the IT department if they would
like to post a complex query to their data-
bases, but with the Optique system, they can
get answers in minutes, he believes.
'This will open up new exploratory and inter-
active ways of working as users get more rel-
evant data sets in shorter time."
The results of the research are planned to be
presented in Hvik, Norway, in early 2015,
with an aim to attract more companies to get
involved, and ultimately develop methods and
technology which will be used by the industry
mainstream.
'We will deliver a good concept, but this will
not be something that can be delivered to the
industry two years from now," Professor
Waaler said.
"I hope that by then [early 2015] we have
something so impressive that the industry will
want to continue to fund this project. I am op-
timistic."
"Optique" $17.5m EU oil and gas big data researchproject now mid-way
View David Holmes talk on video athttp://www.digitalenergyjournal.com/video/1224.aspx
-
Nort
SectorNorwegian,h Sea
platformOil
Simply when you need
d the best.
o eg a,
Tested to the limits.Ensuring ultimate reliacommunications at se
Weve go
ability ofea.
ot
you coveWe ve go
ered.ot
ritmawo
uora
s important to choose the right mas oceans itAs you strive to explore more of the worldIntellian have proved time and time again its high perform.satellite communications partner
wthe to connected you eepketo upon relied be can systems FleetBroadband and TV Sat ,TT,VSAAT.even in the most extreme locations
continuIntellian use:to simple and last to engineered perform,to built ,designsative Innov Seamless integra.produces patented technology to deliver performance leading RF capability
e Intellian the preferred choiceand superior remote management and control capability makkes leading service providersthe world
trust Intellian.When the bottom line relies on staying connected,
timeanceorld,
uslytion
s first auto switching 2.4 meter antennahe worldTv240M -u-band.system for seamless global operation on C and Ktion
e for
intelliantech.com
-
Sharecat of Norway is helping oil and gas
companies put together structured informa-
tion about all of the parts and components in
their offshore operations faster, by sharing
the data.
The basic idea is that Sharecat maintains its
own databases of standard information about
millions of different parts. If you (as a cus-
tomer) have a certain part on your platform,
Sharecat might already have all the necessary
specifications and materials lists (items you
might need to purchase for it). So you can
use Sharecats data rather than collect it
again from scratch.
In a greenfield project, all suppliers, EPC
contractors and the end client will upload
data and documents, review and collaborate
in the same environment.
Sharecat has been gathering its equipment
catalogue for over 20 years and now has mil-
lions of pieces of data, said Sturle Drageset,
sales and marketing director with ShareCat,
speaking at the Digital Energy Journal Ab-
erdeen conference on November 25, Doing
More with Offshore Engineering Data.
Sharecat is based in Bergen, with offices in
Aberdeen, London and Houston, and a re-
seller agreement with WorleyParsons in Aus-
tralia.
The database can be used both in greenfield
projects, to help put together better part data-
bases faster, and in brownfield projects, to
help operators clean their data up. It also pro-
vides associated consulting services.
Sharecat has global agreements with several
oil majors, to put together a global catalogue
of suppliers for Shell, and run a central infor-
mation store for BP. It manages materials
data for Statoil.
Sharecat has a frame agreement with engi-
neering company WorleyParsons, where
Sharecat will provide specialist information
management software and services to Wor-
leyParsons to sell as part of its service to cus-
tomers.
The services are made available over the
cloud as Software as a Service (SaaS).
Data problems
It is too common for operators not to have
accurate records of what they have installed
on their rigs, he said.
We see inspection reports stating missing
certificates, missing history, impossible to
trace parts in the material master, and miss-
ing documentation and numbering informa-
tion.
This might be expected, when you consider
that it is common for operators to take deliv-
ery of an oil rig or FPSO, but they do not re-
ceive the associated documents, part
information and materials lists, in a usable
format.
The problem can be ignored until it is time to
do modifications, but at this point, engineers
can spend 2-3 hours gathering specification
data (tags) for each component, before they
can put together a purchase order for new
materials.
Every time any information is missing, you
need to search for information and involve
colleagues, which take up hours of expensive
time. Poor information will lead to incorrect
purchase, increasing the cost even more.
On greenfield projects, it should be much
easier to gather necessary information.
But operators are often overwhelmed by the
amount of data. The operator may deal with
only a small number of contractors directly,
but each contractor will go on to send hun-
dreds (or thousands) of purchase orders to
their suppliers and manufacturers, which
generates an enormous amount of documen-
tation.
Sometimes documents for a single compo-
nent (such as a motor) will be sent back to
the operator many times, because this com-
ponent is used as part of many different
pieces of equipment.
The information is sent by email, which
means it easily gets lost, and no-one is sure if
it is correct.
Meanwhile, the operators projects staff do
not necessarily have an incentive to make
sure that the data is good, because after the
project they will move onto a new project,
and leave the data problem to the companys
operations staff.
It can take many thousands man hours to put
together a complete parts database for a new
offshore asset, he said, so it is not a surprise
that the work is often not done, and the proj-
ect is handed to operations staff with only 30
per cent of information available.
We end up with poor and missing informa-
tion in the material master, he said.
As the industry has to lower costs, the EPC
(Engineering Procurement and Construction)
companies must work more efficient and at
the same time deliver better quality. This is
what Sharecats products are tailored for, Mr
Drageset said.
We also experience drilling companies to
struggle with procedures and routines to
build up information correctly, he said. A lot
of data is only entered as free text, and no
possibility to retrieve crucial information in
later maintenance, modifications and pur-
chase of equipment and parts.
Sharecat
Sharecats service is to reduce the overall
workload, by maintaining and continually
updating the shared catalogue of part data
and deliver quality data to the clients.
So for example data about a part such as a
specific ABB motor, which might be used in
thousands of different pieces equipment on
many different offshore assets, information
only needs to be entered once and re-used
many times.
Sharecat provides templates which can be
given to engineering contractors so they
know what data they need to provide, and
which can be automatically uploaded into
Drilling and production technology
16 digital energy journal - February / March 2015
Sharecat - a faster way to make offshore databasesSharecat is helping companies put together databases of technical information much faster, with global agreements with BP,Shell and Statoil
Helping you put together offshore databases faster Sturle Drageset, sales and marketing manager,ShareCat
-
our SeEnable YYo
eed of adoption.
ourmaline OilTotal Teine Energy Tecpetrol (US)T
North AQATalley Energy amarack VVaT
alisman Energy TSwift Energy Surge Energy
advocacyoactive
ed the trust and
vocating strong
on-profit
plier
Supply Chain
ce costs and complexity while increasing the sp
Quantum Resource Resources QEP
Progress Energy Post Rock Energy Pioneer Natural ResourcesPinecrest Energy Peyto Exploration PetroQuest Energy Petro-Hunt
Murphy Exploration & Production Mewbourne Oil Matador Production Marathon MacPherson EnergyLong Run Exploration Linn Operating Legado Resources Legacy Reserves
Oil atural Resources od Energy Energy E&P
fshores Offfobil
Resources rgy E&Pesources
proour through this do e W. industrygas and
has gainOFS Portalntiality of electronic data, . In addition to advt approaches for the industry
e have a nocost. Wreducing and ng eCommerce
organization which consists of diverse supp
approach toward best practices to reduc
Our Members:
Our Community:
ates PetroleumYYaO Energy XT
WPX Energy Windsor Energy Wildhorse Resources Whiting Petroleum
alley Energy est VWermilion Energy Venoco Velvet Energy Vantage EnergyVVaanquard Natural ResourcesVVa
Urban O&G Ultra Resources
win Butte Energy Tundra Oil & Gas Tug Hill Operating Trioil Resources Trilogy Energy Trey ResourcesT
Sure Energy Stone EnergyStatoil Southwestern Energy SM Energy Sinopec Daylight Energy Sheridan Production Shell Sandridge Energy Samson Sabine Oil & Gas Rosetta Resources Rock Energy Rio Oil & Gas Rife Resources Rex Energy Resolute Natural Resources Repsol Regent ResourcesQuicksilver ResourcesQ
Petrobras ResourcesPetrobank Energy & Perpetual Energy
irgina Penn VPengrowth PDC EnergyParsley EnergyEnergy Pacesetter Directional Drilling OxyOMERS Energy Oasis Petroleum
fshore Group Northstar OfffNorthern Blizzard Resources Noble Energy Nexen
Nearburg Producing
Resources NALp y p
Le Norman Operating Laredo Petroleum Jones Energy Jetta Operating
Indigo Minerals Husky Oil Operations Huntington Energy Hunt Oil Hilcorp EnergyHighmount Exploration & ProductionHess Corp Harvest Operations Halcon Resources Grizzly Oil Sands
estern O&G Great WGeoSouthern Energy Gear Energy Gastar Exploration USAFreeport - McMoRan O&G Forest O
Na