Newer
Older
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
# coding=utf-8
from bscearth.utils.date import parse_date, add_months
from earthdiagnostics.diagnostic import *
from earthdiagnostics.frequency import Frequencies
import iris
import iris.coord_categorisation
import iris.time
class DaysOverPercentile(Diagnostic):
"""
Calculates the montlhy percentiles
:param data_manager: data management object
:type data_manager: DataManager
:param startdate: startdate
:type startdate: str
:param member: member number
:type member: int
:param chunk: chunk's number
:type chunk: int
:param variable: variable to average
:type variable: str
"""
alias = 'daysover'
"Diagnostic alias for the configuration file"
def __init__(self, data_manager, startdate, member, chunk, domain, variable, leadtime, percentile):
Diagnostic.__init__(self, data_manager)
self.startdate = startdate
self.member = member
self.chunk = chunk
self.variable = variable
self.domain = domain
self.leadtime = leadtime
self.percentile = percentile
def __eq__(self, other):
return self.startdate == other.startdate and self.member == other.member and self.chunk == other.chunk and \
self.domain == other.domain and self.variable == other.variable and self.leadtime == other.leadtime and \
self.percentile == other.percentile
def __str__(self):
return 'Days over percentile Startdate: {0} Member: {1} Chunk: {2} ' \
'Variable: {3}:{4} Leadtime: {5}'.format(self.startdate, self.member, self.chunk, self.domain,
self.variable, self.leadtime)
@classmethod
def generate_jobs(cls, diags, options):
"""
Creates a job for each chunk to compute the diagnostic
:param diags: Diagnostics manager class
:type diags: Diags
:param options: domain, variable, percentil number, maximum depth (level)
:type options: list[str]
:return:
"""
options_available = (DiagnosticDomainOption(),
DiagnosticOption('variable'),
DiagnosticListIntOption('leadtimes'),
DiagnosticFloatOption('percentile'))
options = cls.process_options(options, options_available)
job_list = list()
for startdate, member, chunk in diags.config.experiment.get_chunk_list():
for leadtime in options['leadtimes']:
job_list.append(DaysOverPercentile(diags.data_manager, startdate, member, chunk,
options['domain'], options['variable'], leadtime,
options['percentile']))
return job_list
def request_data(self):
var_name = self.variable + 'prct'
self.percentiles_file = self.request_chunk(self.domain, var_name, None, None, None,
frequency=Frequencies.climatology)
self.variable_file = self.request_chunk(self.domain, self.variable, self.startdate, self.member, self.chunk)
def declare_data_generated(self):
var_name = self.variable + '_daysover'.format(self.leadtime)
self.percentiles_file = self.declare_chunk(self.domain, var_name, None, None, None,
frequency=Frequencies.climatology, vartype=VariableType.STATISTIC)
def compute(self):
"""
Runs the diagnostic
"""
percentiles = iris.load_cube(self.percentiles_file.local_file)
percentile = percentiles.extract(iris.Constraint(percentile=self.percentile))
var = iris.load_cube(self.variable_file.local_file)
date = add_months(parse_date(self.startdate), self.leadtime, 'standard')
leadtime = iris.time.PartialDateTime(date.year, date.month)
iris.coord_categorisation.add_categorised_coord(var, 'leadtime', 'time',
lambda coord, x: coord.units.num2date(x) == leadtime)
var = var.extract(iris.Constraint(leadtime=True))
days_over = var > percentile
print(days_over)