E-Government Portal Updates` Evaluation:

54 International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015
E-Government Portal
Updates’ Evaluation:
A Comparative Analysis
Leonidas Anthopoulos, Business School, Technological Education Institute of Thessaly,
Thessaly, Greece
Kleanthis Sirakoulis, Business School, Technological Education Institute of Thessaly,
Thessaly, Greece
ABSTRACT
More than a decade has passed since the launch of the initial e-Government one-stop web portals, which
concern central points for digital access by citizens, enterprises and government. Due to the broad audience
that these portals serve, various analyses have been performed concerning their effectiveness with regard to
service delivery; trustworthiness with regard to service availability; usability; accessibility; and user satisfaction etc. The results from these analyses have extreme interest for governments, since they reflect government
strategic performance, internal efficiency and effectiveness, while they have been utilized for their upgrades.
E-Government portal upgrade appears to be something usual and various updates have been observed in most
portals during this timeframe. This paper addresses and important issue: “do e-Government portal updates
enhance user satisfaction?” To this end, a comparative qualitative evaluation of some major e-Government
portals is performed, with the use of the American Customer Satisfaction Index (ACSI) during 2009 and 2012.
Keywords:
Accessibility, e-Government, Evaluation, One-Stop, Portal, Usability, User Satisfaction
1. INTRODUCTION
E-Government has been introduced since the
early 1990s as the means to transform government processes to more effective, efficient and
transparent ones; to engage citizens in policy
and decision making; and to modernize public
processes in general (Anthopoulos & Fitsilis,
forthcoming). This government transformation
has been based on various initiatives, according
to a corresponding e-strategic planning, which
concern –among others- investments on information and communication technologies (ICT);
civil servants’ training on ICT skills; diffusion
programs such as, campaigns; public process
reengineering; and respective legislature’s
alignment etc. All these efforts result in citizen
points of access that deliver public information
and services to target audiences.
Alternative points of government access are
named channels (Janssen et al., 2003; Vasilakis
et al., 2007; Reddick & Turner, 2012) and con-
DOI: 10.4018/ijpada.2015040104
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015 55
cern web portals, call centers, traditional office
visits etc. Channels’ performance reflects public
sector’s efficiency and effectiveness, since service and information delivery is the outcome
of the public internal processes. Effectiveness
deals with the quality of public services and
the degree of its independence from political
willing; the quality of policy formulation and
implementation; and the credibility of the government’s commitment to such policies (World
Bank, 2012). On the other hand, government
efficiency concerns public sector performance
or productivity rates and it is mainly associated
with public spending effects to socio-economic
indicators (Hauner & Kyobe, 2008). Both these
definitions justify that these points of access
deliver information and service flows according
to the existing public sector’s efficiency and
effectiveness and to this end, their performance
is important to be measured.
This paper focuses on web channels and
more specifically to the one-stop e-Government
portals, which concern single-points, centrally
official websites, from which citizens access
their governments (Wimmer, 2002; Anthopoulos et al., 2007). The existence of one-stop
e-Government portals requires (Wimmer, 2002)
interconnected agencies; service integration;
and content and service presentation in a logical
manner. Such portals are supposed to be unique
for national cases, as well as for supranational
efforts. Indicative representatives concern USA.
gov (for the U.S. Federal Government); Gov.uk
(for the U.K. Government) etc., while youreurope.eu is a representative supranational case.
The evaluation of e-Government portals
and websites has been approached by various
scholars, while e-Government assessment
alone is of great scientific and political interest.
However, the evaluation of one-stop portals’
updates has not been investigated. More specifically, although most of these portals have
been updated –even more than once- from
their initial appearance, which could be justified by corresponding technological evolution,
service integration or usability improvements,
the outcome from these updates has not been
measured. For instance, USA.gov initially ap-
peared as FirstGov (in late 1990s) (Thompson
et al., 2003) and it was updated to at least 3
versions until today.
With regard to portals’ updates, this paper
aims to answer the following questions: “how
can end-user satisfaction from one-stop eGovernment web portals be measured?” and
“do e-Government portal updates enhance
user satisfaction?”. These two questions are
crucial to be answered for e-Government
scholars. More specifically, the answer to the
first question will explore existing satisfaction
measurement methods from e-Government
portals. On the other hand, the answer to the
second question will demonstrate whether onestop portals’ updates succeed in their mission,
which should be the end-user satisfaction against
previous versions.
The remaining of this paper is structured as
follows: section 2 presents the context of onestop e-government portals and compares user
satisfaction methods. In section 3 the research
method and the corresponding domain study
are illustrated, while the extracted findings are
summarized and discussed. Finally, section 4
contains results and future thoughts.
2. BACKGROUND
This section presents the theoretical context
of one-stop e-Government portals and e-Government satisfaction. One-stop e-Government
concerns a single point of access to public
information and services offered even-by different public authorities (Wimmer, 2002). The
existence of online one-stops requires agencies’
interconnection, integration and interoperability -mainly based on open standards-, while
end-users (citizens, enterprises or other public
agencies) are must hold the appropriate skills
and infrastructure to access these single points
of access (Borras, 2004). From their appearance
in late 1990s one-stop portals were primary
based on e-commerce technologies in order to
deliver online services via web and voice channels to e-Government end-users (Anthopoulos
et al., 2007). Most of them adopted usability
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
56 International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015
Table 1. e-Government web portal evaluation frameworks/models
e-Government web portal evaluation framework/
model
Scholars
W3C WAI
Taufik et al. (2007)
Usability Assessment Framework
Nielsen (1993); Silius & Tervakari (2003); Large et al.
(2002); Glassey & Glassey (2004); Wilder (2007)
Usability; Content consistency; and Openness
Kokkinari et al. (2005)
Web Diagnostic Tools
Choudrie et al. (2004)
TAM and Section 508 Guidelines
Jaeger & Matterson (2009)
Formative and Summative
Thompson et al. (2003)
Five-index evaluation model
Garcia et al. (2005)
(Wilder, 2007) and accessibility (Taufik et al.,
2007) features early, as a means to enhance
end-users’ satisfaction (Zhang & von Dran,
2000). These features vary from World Wide
Web Consortium’s (W3C) Web Accessibility
Initiative (WAI) standards to even localization
options in terms of cultural and social trends and
habits (del Rio, 2013). To this end, Reddick &
Turner (2012) determine that user satisfaction
impacts service channel selection and concluded
that e-Government rather becomes an alternative –to traditional channels- mean for citizens
to access their governments, instead of a global
change for public transactions.
The evaluation of e-Government portals
(Table 1) can be performed with means of
usability’s evaluation, basically presented by
Nielsen (1993) and later adopted by other
scholars for web portal evaluation (i.e., Silius
& Tervakari (2003); Large et al. (2002); Glassey
& Glassey (2004)). Kokkinari et al. (2005) extended Nielsen’s framework in order to evaluate
a set of Cyprian e-Government portals. In their
analysis, they considered except from usability
features others related to content’s consistency
and simplicity; and openness to the public.
Choudrie et al. (2004) used web diagnostic tools
to evaluate e-Government web sites. Jaeger
& Matteson (2009) evaluated e-Government
website with the examination of the Technology
Acceptance Model (TAM) and the adoption of
Section 508 Guidelines. Thompson et al. (2003)
combined all the above techniques and per-
formed a normative and summative evaluation
of federal e-Government web portals. Garcia et
al. (2005) introduced a five-index model, which
evaluates e-Government web portals according
to cognitive effort; tolerance; reach; physical
effort; and trust.
With regard to satisfaction’s context, the
Expectation-Confirmation Theory (ECT) (Oliver, 1980) states that consumers ground their
initial expectation -prior to purchase- and after a
period of initial consumption they build perceptions about the performance of the consumed
product/service. Next, consumers will decide
on their level of satisfaction based on the extent
to which their expectation is confirmed through
comparing the actual performance of the product/service against their initial expectation of
the performance (Alawneh et al., 2013). Satisfied consumers form re-purchasing intentions,
which is called consumer loyalty.
Beyond the abovementioned technical and
usability criteria, authors performed a literature
review with regard to e-Government satisfaction
measurement and identified various existing
frameworks and models (Table 2). One-stop
e-Government portals’ performance and corresponding satisfaction has been investigated
by various scholars and organizations so far
and alternative models can be located (Fitsilis
et al., 2010). These models can be aggregated
in satisfaction’s direct and indirect measurements. Direct for instance, are utilized by Cap
Gemini, which has performed annual investiga-
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015 57
Table 2. Literature review on e-Government satisfaction measurement
e-Government Evaluation
Framework/Model
Direct/Indirect Satisfaction
Measurement
Scholars/Organization
e-Government project success
evaluation framework
Indirect
Fitsilis et al. (2010)
e-Government benchmark
framework
Direct
Capgemini (2011)
American Customer Satisfaction
Index (ACSI)
Direct
Accenture (2012), Park (2007)
e-Service Quality
Indirect
Magoutas et al. (2007);
Papadomichelaki & Mentzas (2012)
e-Government Readiness
Indirect
Tucker (2012), United Nations
(2012)
COBRA
Indirect
Osman et al. (2011)
e-Government Acceptance Model
Indirect
Park (2007)
e-Satisfaction Determinants
Direct
Alawneh et al. (2013)
tions since 2001, with regard to online availability of public services across Europe (Cap
Gemini, 2011). Cap Gemini has introduced an
e-Government benchmark framework, which
aims to measure user satisfaction (Cap Gemini,
2012), via comparing user experience from
public service execution, with experiences
from non-governmental services (i.e., banking
services). Moreover, Accenture (2009; 2012)
has performed various investigations concerning citizen satisfaction from e-Government
services. These measurements have been based
on the American Customer Satisfaction Index
(ACSI) and recognized early dissatisfaction in
2005 (when citizens turned back to traditional
channels), while today digital citizens appear
to demand more from their governments:
findings from U.S.A. show that the majority
of people would use digital services if offered
by government, especially for routine transactions. And over half want to conduct all their
government business digitally in the future
(Accenture, 2012).
On the other hand, indirect measurements
focus on service availability and quality. Magoutas et al. (2007) for instance, introduced an
e-service quality measurement model; Papadomichelaki & Mentzas (2012) measured e-service
quality too, with a multiple-item scale model;
Tucker (2012) compared several assessment
instruments (i.e., SERVQUAL, governance,
Strategic Alignment Model etc.) and concluded
on an e-Government readiness model, which
measures organization’s efficiency for e-service
delivery. This approach has similarities to the
United Nations (UN) e-Government Readiness
Index (EGDI) (United Nations, 2012), which
measures respective public sector’s internal
capacity. Readiness lies behind one-stop portal’s
performance since it corresponds to service
availability and quality. However, EGDI is
analyzed in three other indexes, one of which is
the Online Service Index, which measures service availability and accessibility. Osman et al.
(2011) compared various e-government Value
Measurement Models (VMM), e-Government
success models and e-Government Service
Quality models and introduced their individual
evaluation framework, which associates satisfaction with Costs, Opportunities, Benefits
and Risks Analysis. Park (2007) compared
the introduced an e-Government acceptance
model, which was based on the combination
of three widely accepted theoretical evaluation
models (SERVQUAL, ACSI and Technology
Acceptance Model (TAM)). Alawneh et al.
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
58 International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015
(2013) validated a set of determinants, which
influence user satisfaction from e-Services:
security and privacy; trust; accessibility; awareness; and quality.
All the above findings give answer to
the first question of this article with regard to
“how can end-user satisfaction from one-stop
e-Government web portals be measured”. More
specifically, literature findings depict that enduser satisfaction can be measured either directly
with qualitative studies or via service quality
evaluation; or indirectly with the assessment of
technical features, web metrics, service quality
or acceptance models. Although this literature
review was not exhaustive, it returns a broad
picture concerning how different schools
of thought propose methods for one-stop eGovernment evaluation.
3. DOMAIN STUDY
In order to answer the second question of this
paper, this section contains qualitative findings
from the evaluation of the updates of some major
one-stop e-government portals. The selection
of the examined portals occurred in late 2009
for the purposes of the module “Graduate
Seminar”, which is performed at the Project
Management Department of the Technological
Education Institute (TEI) of Thessaly, Greece.
Only students who have fulfilled their studies
and are close to their graduation can attend
this module, while the context of this module
concerned e-Government Strategic and Project
Management during the academic years from
2009 to 2012. Among the other teaching context,
students investigated and became aware of the
Greek one-stop e-Government portal, named
ERMIS (www.ermis.gov.gr), as a means to
define initial expectations from a corresponding website and to evaluate their satisfaction
accordingly.
3.1. Research Methodology
A sample of a total amount of 256 students
was participated during the period lasted from
October 2009 to June 2012 in a qualitative
survey over their satisfaction from different
one-stop e-Government portals. This sample
was homogeneous, consisting of young citizens
all aged between 22 and 24, all completing
their graduate educational level; they were all
good English speakers, which was the primary
language in all the examined cases, except from
the German case, which is offered in English too.
However, even the German case was examined
in the German language for the corresponding
e-service execution, with the support of German translators. Moreover, all students where
taught about one-stop e-Government under
the Graduate Seminar module and the Greek
respective portal. The participation was obligatory for the purposes of this module, while two
different tutors observed students’ attitudes
during this study in order to secure survey’s
objectivity. In this paper the findings from the
analysis only from the 63 of the participants
is presented, since the overall analysis has not
been completed yet. Thus some reasonable
limitations are introduced, but the outcomes
from this study can be considered efficient to
generate a primary picture over the comparative
satisfaction that the examined websites and their
updates provided.
The selection of the examined one-stop
portals was based on the following criteria: a)
size of target-audience and project size, which
would lead to potential updates (something
that really happened in the examined period);
b) launching year (as early as possible); c)
location, so that representatives only from
developed countries of all continents would be
incorporated; d) ranked in the top 35 countries in
e-Government readiness index, but with different values (United Nations, 2008), meaning that
representatives from top to lower positions had
to be selected. The investigation started during
the winter semester of 2009 and continued on
the same websites until the spring semester of
2012, although their classification according to
the e-Government readiness index changed. The
reason that the selected websites remained the
same was the purpose of this paper, which is to
demonstrate a comparative evaluation of each
website’s updates during the same timeframe.
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015 59
Table 3. sample’s size over time
Year
Participants
2009
12
2010
4
2011
28
2012
19
To this end, the following five (5) portals
(Table 4) were identified to fulfill all these criteria: United States of America (U.S.A.) (USA.
gov), which was launched in 2000 as FirstGov,
changed its name in 2007 and ranked 4th in eGovernment readiness index in 2008; United
Kingdom (U.K.) (gov.uk), which was online in
1999 as UKOnline, renamed to directgov.uk in
2004 and to gov.uk in 2012, while it was rated
10th in e-Government readiness index; Canada
(Canada.gc.ca), which was available online
in 2000 and ranked 7th; Australia (Australia.
au) that was sorted in the 8th position of the
e-Government readiness index and launched in
2000; and Germany (bund.de) that was ranked
22nd and launched in 2000. Moreover, one
supranational case was selected from the European Union, named YourEurope.eu that was
initially online in 2005. Finally, for reasons of
comparison with the previously selected cases,
a smaller national representative was located
in New Zealand (ranked 18th in e-Government
readiness index) and it was selected for the
following reasons: it was implemented entirely
from national funding and efforts, while and it
serves a quite small community compared to
the other selected cases. Various updates were
observed in time for all the examined portals,
some of which were dramatic (i.e., USA.gov
renovation in 2010 and Gov.uk renovation
in 2012), while others concerned small-scale
changes (i.e., corresponding e-service changes).
For the purposes of this study, students
behaved as end-users and measured their satisfaction from the offered online content and
services, during the execution of the same, predefined public service: “locate a postgraduate
program in the examined country in the project
management domain”. The selection of this eservice was based on the following reasons: it
was fully understood by the participants, who
realized its successful execution when they
were able to locate such a program. Moreover,
it is a service that is offered to foreigners in
all the examined cases, since most one-stop
e-Government portals focus on locals and immi-
Table 4. the selected one-stop e-Government portals
Country
One-Stop e-Government
Portal
e-Government
Index Ranking
(2008)
e-Government
Index Ranking
(2012)
Launch
Year
USA
www.usa.gov
4th
5th
2000
UK
www.gov.uk
10
3rd
1999
Canada (CAN)
www.canada.gc.ca
7th
11th
2000
Australia (AUS)
www.australia.au
8th
12th
2000
Germany (DE)
www.bund.de
22
17th
2000
European Union (EU)
www.youreurope.eu
0.59 (value)
0.72 (value)
2005
New Zealand (NZL)
www.newzealand.govt.nz
18th
13th
2002
th
nd
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
60 International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015
Figure 1. The American Customer Satisfaction Index (ACSI)
grants and the corresponding e-services cannot
be tested by someone who is not aware of the
particular legislature and internal environment
of a foreign country.
Each student evaluated individually its
satisfaction, with regards to the expectation
from the Greek ERMIS one-stop portal. Different students each semester faced potentially
different versions of the examined websites,
which was one of the requirements of this
study, without being aware of the previous
version –which secures the objectivity of this
study-. The satisfaction measurement has been
utilized aimed to evaluate the updates of the
examined portals with the end user satisfaction
measurement.
Satisfaction was measured with the use of
the ACSI due to its direct satisfaction measurement and simplicity among the other theoretical
models, which were presented in (Table 2). The
ACSI consists of the following elements (Figure
1) (Park, 2007):
•
•
Customer expectations’ determination,
which concerns anticipation prior to consumer’s experience;
Perceived quality, which represents customer evaluation of the quality of his recent
consumption experiences;
•
•
•
Perceived value that defines the pricerelative quality;
Customer complaints that document individual comments regarding the experience
from the consumed product/service;
Customer Loyalty, which demonstrates
whether the customer intends to consume
this product/service again.
For the purposes of the ACSI, all participants were requested to answer the following
questions with the Likert scale from 0 to 10 (0:
not satisfied; to 10: very satisfied). The scale
was selected to be broad in order to identify
even small variances between qualitative results.
Q1: How easily can I understand where I am?
The purpose of this question was to examine
usability features, which were previously
explained to the participants (i.e., brand
name, logo size and position, explanation
text and figures etc.). This question was
crucial to develop an initial satisfaction
feeling of the participant, since none was
previously aware of the examined case that
was presented to him.
Q2: How much satisfied am I from the online
service? This question was answered after
the execution of the abovementioned on-
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015 61
Table 5. Results’ change during time
Q1
Q2
Q3
Q4
YEAR
N
Mean Rank
Mean Rank
Mean Rank
Mean Rank
2009
71
147,2
136,78
94,54
101,96
2010
20
203,15
158,4
162,13
121,58
2011
107
142,89
154,47
172,6
176,42
2012
125
180,18
183,35
191,22
190,23
Total
323
line service. The overall satisfaction was
generated by the successful location of the
e-service’s target (the identification of an
appropriate to the participant postgraduate
program in project management) and it was
evaluated with Likert values.
Q3: How easily can I find what I look for? This
question was answered after the execution
of the abovementioned online service. Students defined their satisfaction with Likert
values, regarding how easily he navigated
in the website, found the appropriate path
to the objective, as well as concerning the
existence other usability features (i.e., help,
wizards, search etc.).
Q4: Would I use this website for the execution
of the same or similar e-service? This
question aimed to measure with Likert
values the customer’s loyalty concerning
the examined website and it is a prerequisite
for the ACSI.
3.2. Statistical Analysis
Collected data were analyzed with the SPSS
version 17 software, which is appropriate and
widely accepted for statistical analyses. The first
outcome from this analysis illustrates that participants’ answers improved during time (Table
5) (Figure 2), which can be interpreted to an
overall improvement of their gained experience
from the examined cases. This improvement
is statistically important according to Kruskal
Wallis test (Table 6).
With regard to the analysis of each question’s individual results, Kruskal Wallis tests
Figure 2. Sample’s overall experience over time
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
62 International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015
Table 6. Kruskal Wallis test’s results regarding sample’s answers over time
Q1
Q2
Q3
Q4
Chi-Square
15,322
12,742
51,864
47,888
Df
3
3
3
3
Asymp. Sig.
,002
,005
,000
,000
Chi-Square*
11,189
12,701
51,288
43,106
df*
2
2
2
2
Asymp. Sig.*
,004
,002
,000
,000
illustrate significant changes in time (Table 7)
for all the questions. YourEurope portal received
good ranking steadily in question 1 (Q1), while
German portal gathered low ranking in the same
question. Moreover, Canadian portal’s evaluation in Q1 significantly rises in 2012, which can
be interpreted to a significant improvement in
website’s corresponding appearance (Figure 3).
The answers in question 2 (Q2) show
similar performance of the European and the
German portals, while New Zealand’s perfor-
mance raises significantly during 2012 (Figure
4). Moreover, users ranked steadily high YourEurope portal -except from 2010- and steadily
low the German website in questions 3 (Q3) and
4 (Q4), while the Canadian performance rises
importantly after 2010 in the same question
(Figures 5, 6). All these findings demonstrate
an overall sample’s satisfaction from European
portal, dissatisfaction from the German case and
variances in time for the other examined cases.
Table 7. Kruskal Wallis test’s results regarding individual question’s answers over time
2009
2010
2011
2012
Chi-Square
25,371
10,621
67,141
18,492
df
5
4
4
6
Asymp. Sig.
0
0,031
0
0,005
Chi-Square
35,845
14,947
66,72
38,889
df
5
4
4
6
Asymp. Sig.
0
0,005
0
0
Chi-Square
28,427
16,028
50,914
41,554
df
5
4
4
6
Asymp. Sig.
0
0,003
0
0
Chi-Square
35,71
14,55
43,417
32,483
df
5
4
4
6
Asymp. Sig.
0
0,006
0
0
Q1
Q2
Q3
Q4
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015 63
Figure 3. Q1 performance in time
Then, each examined case was analyzed
with regard to participants’ replies in time
(Table 8). To this end, USA.gov (Figure 7) did
not illustrate important differences in time for
the Q1, which shows that despite this portal’s
updates, participants do not appear to be satisfied from its usability features. Kruskal Wallis
analysis does not return significant differences
for U.K. (Figure 8), which shows that participants gained the same experience in time
from Gov.uk portal. Canadian portal (Figure
9) shows significant variances in time for all
the questions, which shows that this portal’s
updates affected sometimes positively and
others negatively user satisfaction. Data for
YourEurope portal (Figure 10) shows that Q1
and Q2 were not affected in time, meaning that
users had the same experience from its usability
features and expectations; Q3 and Q4 were increased in time, which determines a significant
improvement in overall satisfaction and loyalty
from the participants. Australian one-stop portal
(Figure 11) shows significant variances in time
for all the questions, but the overall opinion was
Figure 4. Q2 performance in time
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
64 International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015
Figure 5. Q3 performance in time
evolved in time, which shows that this portal’s
updates affected positively user satisfaction.
Users’ rankings for Bund.de (Figure 12) return
unimportant variances for all the questions but
Q1; this portal’s updates affected negatively
user satisfaction from usability features in 2011.
Finally, findings from New Zealand (Figure 13)
depict an important growth in time, with regard
to user satisfaction.
In order to identify significant changes from
year to year, a Mann – Whitney and Wilcoxon
analyses was performed on the collected data
(Table 9). This analysis does not return statistically important variances between 2009 and
2011 for Q1, but significant changes for the other
questions. This finding can be interpreted to
the following: although usability features were
updated due to various renovations (i.e., USA.
gov redesign in 2010) between 2009 and 2011
in the examined portals, the overall satisfaction’s experience from the tested e-service was
not affected. On the other hand, no significant
change appears for the period 2011 and 2012.
Figure 6. Q4 performance in time
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015 65
Table 8. Kruskal Wallis test’s results regarding each portal’s performance over time
USA
N
Q1
Q2
Q3
Q4
2009
12
25,58
16,58
13,92
9,63
2010
4
42,38
24
34,75
31,75
2011
28
36,16
34,29
28,93
35,68
2012
19
27,74
40,05
47,37
40,76
Chi-Square
5,579
14,143
26,536
23,777
df
3
3
3
3
Asymp. Sig.
0,134
0,003
0
0
UK
N
Q1
Q2
Q3
Q4
2009
11
21,14
14,86
11,91
13,5
2010
4
30,63
34,13
30,25
31,63
2011
20
28,25
31,53
36,6
33,95
2012
19
29,74
29,18
26,37
27,95
Chi-Square
2,495
9,721
18,894
12,695
df
3
3
3
3
Asymp. Sig.
0,476
0,021
0
0,005
Canada
N
Q1
Q2
Q3
Q4
2009
12
27,33
31,25
12,5
11,83
2010
4
42,13
30
20
25,25
2011
19
12,5
17,24
36,89
32,32
2012
19
39,53
34,87
29,16
33,05
Chi-Square
33,273
13,316
19,181
16,436
df
3
3
3
3
Asymp. Sig.
0
0,004
0
0,001
N
Q1
Q2
Q3
Q4
2009
12
27,04
18,88
16,29
27,96
2010
4
42,88
33,5
40,13
2,63
2011
20
30,43
33,2
24,5
28,08
2012
19
YourEurope.eu
22,92
27,13
36,53
33,29
Chi-Square
6,294
7,449
15,914
12,869
df
3
3
3
3
Asymp. Sig.
0,098
0,059
0,001
0,005
N
Q1
Q2
Q3
Q4
12
12,13
11,04
7,96
9,17
Australia
2009
continued on following page
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
66 International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015
Table 8. Continued
2010
2011
2012
19
18,45
19,13
21,08
20,32
Chi-Square
3,898
6,65
16,433
11,622
df
1
1
1
1
Asymp. Sig.
0,048
0,01
0
0,001
N
Q1
Q2
Q3
Q4
2010
4
26,75
28,75
17,63
16,38
2011
20
14,35
18,78
20,68
19,8
2012
19
29,05
23,97
24,32
25,5
Chi-Square
14,248
3,018
1,414
2,959
df
2
2
2
2
Asymp. Sig.
0,001
0,221
0,493
0,228
N
Q1
Q2
Q3
Q4
12
7,88
6,75
7,46
6,83
11
Australia
2009
Germany
2009
2010
2011
2012
16,5
17,73
16,95
17,64
Chi-Square
9,502
15,517
11,501
14,899
df
1
1
1
1
Asymp. Sig.
0,002
0
0,001
0
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015 67
Figure 7. USA.gov
The same analysis for each of the examined case (Table 10) identifies the following:
a significant improvement in Q3 appears for
the USA, meaning that the examined service
was improved during the examined period; Q3
becomes worse in time for the UK portal, which
shows that the same service was not improved
but rather declined according to the participants; Q1 and Q2 were improved for Canada,
which determine a usability improvement and
an overall satisfaction increment with regard
to participants’ expectations; Q2 improves in
2009-2011 for the EU portal, which confirms
a corresponding improvement in the examined
e-service; finally, Q3 rises during 2011 to 2012
for the EU portal, which depicts an overall satisfaction increment compared to respondents’
expectations.
3.3. Discussion
The above findings come out only from the
24.6% (63 out of 256) of the participants of
this study, since the analysis is still in progress.
Figure 8. Gov.uk
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
68 International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015
Figure 9. Canada.gc.ca
However, the outcomes can be considered
efficient to provide a primary picture of this
comparative study, due to the sample’s homogeneity and survey’s objectivity. Moreover,
participants were fully aware of the context of
one-stop e-Government, while they had developed their expectations from various e-services’
execution in Greek one-stop e-Government
portal and from the corresponding e-service
execution in Greece.
The qualitative data that have been collected were examined with various tests with
the SPSS version 17 and the following findings
are returned:
•
•
Participants’ answers were improved
during time, which can be interpreted to
an overall improvement of their gained
experience from the examined cases.
Participants appear more satisfied in general from YourEurope.eu portal and this
Figure 10. YourEurope.eu
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015 69
Figure 11. Australia.au
•
•
satisfaction increases during the examined
period and transforms to an increased
loyalty.
Students feel satisfied from YourEurope’s
usability features and steadily dissatisfied
from the German portal’s usability. On the
contrary, their corresponding satisfaction
increased in 2012 for the Canadian case,
which is interpreted to a significant website’s update.
USA.gov was renovated in 2010. However,
this update did not affect participants’ feelings against this portal’s usability.
•
•
Gov.uk was updated significantly with
regard to its appearance in 2012. However,
participants were not satisfied from its usability renovation. In general, the answers
from all the questions do not generate
important variances in participants’ satisfaction from the corresponding updates.
Participants’ satisfaction from Australian
portal varies in time for all the questions,
but the overall opinion was evolved, which
shows that this portal’s updates affected
positively user satisfaction.
Figure 12. Bund.de
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
70 International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015
Figure 13. newzealand.govt.nz
•
Findings from New Zealand depict an
important growth in time, with regard to
user satisfaction.
All these findings show that not all updates
and renovations focus on customer satisfaction
neither they succeed in meeting user expectations. Moreover, the complexity of one-stop
portals’ evaluation is validated. For instance,
corresponding developers must be aware of
all the potential audiences and leverage their
satisfaction. Furthermore, all the examined
cases redirected transparently the user to a
corresponding educational website to fulfill
this e-service and locate the postgraduate program, which means that the examined e-service
concerns a four-stage one (Anthopoulos et al.,
2007). However, none of the respondents realized this redirection and the overall satisfaction
was connected to the initially accessed one-stop
portal. This final finding is extremely important
for one-stop portal developers and opens new
research areas, such as “connected web experiences” or “interconnected satisfaction”.
Table 9. Mann: Whitney and Wilcoxon analyses for 2-year comparisons
2009-2011
Q1
Q2
Q3
Q4
Mann-Whitney U
1924,000
1574,500
879,000
863,000
Wilcoxon W
3052,000
2702,500
2007,000
1991,000
Z
-,574
-2,224
-5,507
-5,570
Asymp. Sig. (2-tailed)
,566
,026
,000
,000
2011-2012
Q1
Q2
Q3
Q4
Mann-Whitney U
2843,500
3005,000
2938,500
3044,000
Wilcoxon W
6671,500
6833,000
6766,500
6872,000
Z
-1,562
-1,016
-1,251
-,889
Asymp. Sig. (2-tailed)
,118
,310
,211
,374
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015 71
Table 10. Mann: Whitney and Wilcoxon analyses for 2-year comparisons for each case
USA
Q1
Q2
Q3
Q4
Mann-Whitney U
203,000
210,500
109,000
216,000
Wilcoxon W
393,000
616,500
515,000
622,000
Z
-1,406
-1,255
-3,466
-1,103
Asymp. Sig. (2-tailed)
,160
,209
,001
,270
UK
Q1
Q2
Q3
Q4
Mann-Whitney U
178,000
180,500
118,500
151,500
Wilcoxon W
388,000
370,500
308,500
341,500
Z
-,344
-,275
-2,104
-1,103
Asymp. Sig. (2-tailed)
,731
,783
,035
,270
Q1
Q2
Q3
Q4
Mann-Whitney U
18,500
66,000
129,500
160,500
Wilcoxon W
208,500
256,000
319,500
350,500
Z
-4,898
-3,365
-1,518
-,593
Asymp. Sig. (2-tailed)
,000
,001
,129
,553
Q1
Q2
Q3
Q4
Mann-Whitney U
101,000
43,500
79,500
117,000
Wilcoxon W
179,000
121,500
157,500
195,000
Z
-,784
-3,335
-1,636
-,121
Asymp. Sig. (2-tailed)
,433
,001
,102
,903
Q1
Q2
Q3
Q4
Mann-Whitney U
141,500
162,500
103,000
148,500
Wilcoxon W
331,500
352,500
313,000
358,500
Z
-1,408
-,829
-2,533
-1,220
Asymp. Sig. (2-tailed)
,159
,407
,011
,222
Canada
EU (2009-2011)
EU 2011 - 2012
4. RESULTS AND
FUTURE THOUGHTS
This paper focused on one-stop e-Government
comparative satisfaction evaluation. More specifically it addressed one-stop portals’ renova-
tions and updates and how these updates affect
end-users satisfaction. To this end, the authors
stated two important research questions. The
first one concerned framework and methods
for e-Government satisfaction measurement
and it was answered with literature findings.
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
72 International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015
More specifically, one-stop e-Government is
evaluated both with technical standards (i.e.,
meeting usability features) and with satisfaction criteria.
The second question focused on whether
portal updates enhance user satisfaction. This
question was investigated with the examination of various major e-Government one-stop
portals, whose selection was based on a set
of objective criteria. Seven one-stop portals
were examined during 2009 and 2012 by a
homogeneous student sample in Greece and
the corresponding findings’ analysis is still in
progress. However, existing results illustrate
important findings, which can be utilized by
one-stop e-Government portals. For instance,
USA.gov and Gov.uk renovations did not affect
respondents’ satisfaction from the examined
e-service.
Some limitations are reasonably located
and have to be considered: the first one concern that this research is in progress and not
all data have been analyzed. This limitation
will be overcome with the completion of this
analysis. However, the presented percentage is
quite efficient (almost 25% of the total sample)
and a primary picture can be observed. Another
limitation concerns that only one e-service was
tested and evaluated. However, the importance
of the examined e-service is extensive, due to
the broad audience of educational services.
Additionally, it is not easy for a foreigner to
evaluate a one-stop portal without executing
local e-services. To this end, future research
will combine satisfaction’s evaluation findings
from both e-services that focus to locals and
foreigners. Finally, new research questions
have been identified from the findings of this
paper and address the evaluation of e-four-stage
e-services, which are executed in interconnected
websites. These questions are expected to be
addressed in future research too.
REFERENCES
Accenture (2009). From e-Government to eGovernance: Using new technologies to strengthen
relationships with citizens [online]. Retrieved, March
2014 from http://akgul.bilkent.edu.tr/egov/Accenture_Institute_Health_Public_Service_From_eGovernment_to_eGovernance.pdf
Accenture (2012). Build It and They Will Come?
The Accenture Digital Citizen Pulse Survey and
the Future of Government Operations [online].
Retrieved, March 2014 from http://www.accenture.
com/SiteCollectionDocuments/PDF/AccentureDigital-Citizen-FullSurvey.pdf
Alawneh, A., Al-Refai, H., & Batiha, K. (2013).
Measuring user satisfaction from e-Government
services: Lessons from Jordan. Government Information Quarterly, 30(3), 277–288. doi:10.1016/j.
giq.2013.03.001
Anthopoulos, L., Siozos, P., & Tsoukalas, I. A. (2007).
Applying Participatory Design and Collaboration
in Digital Public Services for discovering and redesigning e-Government services. Government Information Quarterly, 24(2), 353–376. doi:10.1016/j.
giq.2006.07.018
Borras, J. (2004). International Technical Standards
for e-Government. Electronic. Journal of E-Government, 2(2), 139–146.
Cap Gemini. (2011). eGovernment Benchmark
Framework 2012-2015. Retrieved, March 2014
from http://ec.europa.eu/digital-agenda/sites/digitalagenda/files/eGovernment%20Benchmarking%20
method%20paper%20published%20version_0.pdf
Cap Gemini. (2011). Digitizing Public Services in
Europe: Putting ambition into action. 9th Benchmark
Measurement. Retrieved, March 2014 from https://
ec.europa.eu/digital-agenda/sites/digital-agenda/
files/egov_report.pdf
Choudrie, J., Ghinea, G., & Weerakkody, V. (2004).
Evaluating Global e-Government Sites: A View using Web Diagnostic Tools. Electronic. Journal of
E-Government, 2(2), 105–114.
del Rio, M. S. (2013). Improving the Citizen Experience in the use of Mexican Government Websites.
European Scientific Journal, 3, pp. 38-41, ISSN:
1857 – 7881.
Fitsilis, P., Anthopoulos, L., & Gerogiannis, V. (2010).
An evaluation framework for e-government projects.
In Ch. Reddick (Ed.), Citizens and E-Government:
Evaluating Policy and Management. Hershey, PA:
IGI Global; doi:10.4018/978-1-61520-931-6.ch005
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015 73
Garcia, A. C., Maciel, C., & Pinto, F. B. (2005). A
Quality Inspection Method to Evaluate E-Government Sites. In M. A. Wimmer et al. (Eds.), EGOV
2005, LNCS 3591 (pp. 198–209). Berlin, Heidelberg:
Springer-Verlag. doi:10.1007/11545156_19
Papadomichelaki, X., & Mentzas, G. (2012).
e-GovQual: A multiple-item scale for assessing
e-government service quality. Government Information Quarterly, 29(1), 98–109. doi:10.1016/j.
giq.2011.08.011
Glassey, O., & Glassey, O. F. (2004). Proximity
Indicator for e-Government: The Smallest Number
of Clicks. Journal of E-Government, 1(4).
Park, J. H. (2007). How Citizens Accept e-Government Service: A Comparison of Four Theoretical
Models. The Korean Journal of Policy Studies,
21(2), 143–156.
Hauner, D., & Kyobe, A. (2008). Determinants of
Government Efficiency: International Monetary
Fund (IMF) Working Paper. Retrieved, Dec. 2013
from http://www.imf.org/external/pubs/ft/wp/2008/
wp08228.pdf
Jaeger, P., & Matterson, M. (2009). e-Government
and Technology Acceptance: The Case of the Implementation of Section 508 Guidelines for Websites.
Electronic. Journal of E-Government, 7(1), 87–98.
Janssen, M., Wagenaar, R., & Beerens, J. (2003).
Towards a Flexible ICT-Architecture for MultiChannel E-Government Service Provisioning. In
the proceedings of the 36th Hawaii International
Conference on System Sciences (HICSS), IEEE.
Kokkinari, A. I., Mylonas, S., & Mina, S. (2005).
E-Government Initiatives in Cyprus. In the Proceedings of the eGovernment Workshop ’05 (eGOV05).
Large, A., Beheshti, J., & Rahman, T. (2002). Design
Criteria for Children’s Web Portals: The Users Speak
Out. Journal of the American Society for Information
Science and Technology, 53(2), 79–94. doi:10.1002/
asi.10012
Magoutas, B., Halaris, C., & Mentzas, G. (2007).
An Ontology for the Multi-perspective Evaluation
of Quality in E-Government Services. In M. A. Wimmer, H. J. Scholl, & A. Grönlund (Eds.), EGOV 2007,
LNCS 4656 (pp. 318–329). Springer-Verlag Berlin
Heidelberg. doi:10.1007/978-3-540-74444-3_27
Nielsen, J. (1993). Usability Engineering. Academic
Press.
Oliver, R. (1980). A cognitive model of the antecedents and consequences of satisfaction decisions. JMR,
Journal of Marketing Research, 17(4), 460–469.
doi:10.2307/3150499
Osman, I., Anouze, A. L., Irani, Z., Lee, H., Balci,
A., Medeni, T.D. & Weerakkody, V. (2011). A new
COBRAS Framework to Evaluate e-Government
Services: a citizen centric perspective. In the proceedings of the tGov Workshop 2011 (tGOV11).
Reddick, Ch., & Turner, M. (2012). Channel choice
and public service delivery in Canada: Comparing e-government to traditional service delivery.
Government Information Quarterly, 29(1), 1–11.
doi:10.1016/j.giq.2011.03.005
Silius, T., & Tervakari, M. (2003). An evaluation of
the usefulness of web-based learning environments.
The evaluation tool into the portal of Finnish virtual
university. In the Proceedings of the Conference on
University Networks and E-learning.
Taoufik, I., Kabaili, H., & Kettani, D. (2007). Designing an E-Government Portal Accessible to Illiterate
Citizens. In the proceedings of the ICEGOV2007,
ACM ISSN 978-1-59593-822 -0/07/12
The World Bank. (2012) The Worldwide Governance
Indicators (WGI) [online]. Retrieved, December
2013 from http://info.worldbank.org/governance/
wgi/index.aspx#home
Thompson, K. M., McClure, C. R., & Jaeger, P.
T. (2003). Evaluating federal websites: Improving
e-Government for the people. In J. F. George (Ed.),
Computers in society: Privacy, ethics, and the Internet
(pp. 400–412). Upper Saddle River, NJ: Prentice Hall.
Tucker, S.-P. L. (2012). Assessing and Modeling
the Readiness of Electronic Government. International Journal of Electronic Commerce Studies, 3(2),
251–270. doi:10.7903/ijecs.1094
United Nations. (2008). E-Government Survey
2008: From e-Government to Connected Governance [online]. Retrieved, March 2014 from http://
unpan1.un.org/intradoc/groups/public/documents/
un/unpan028607.pdf
United Nations. (2010). E-Government Survey
2010: Leveraging e-government at a time of financial and economic crisis [online]. Retrieved,
March 2014 from http://www.epractice.eu/files/
UN%20E-Government%20Survey%202010%20
-%20Part%20I.pdf
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.
74 International Journal of Public Administration in the Digital Age, 2(2), 54-74, April-June 2015
United Nations. (2012). E-Government for the People
E-Government Survey 2012 [online]. Retrieved,
March 2014 from http://unpan1.un.org/intradoc/
groups/public/documents/un/unpan048065.pdf
Vasilakis, C., Lepouras, G., & Halatsis, C. (2007).
A knowledge-based approach for developing multichannel e-government services. Electronic Commerce Research and Applications, 6(1), 113–124.
doi:10.1016/j.elerap.2006.07.004
Wimmer, M. (2002). Integrated service modelling
for online one-stop government. Electronic Markets,
12(3), 149–156. doi:10.1080/101967802320245910
Zhang, P., & von Dran, G. M. (2000). Satisfiers
and Dissatisfiers: A two-factor Model for Website
Design and Evaluation. Journal of the American
Society for Information Science, 51(14), 1253–1268.
doi:10.1002/1097-4571(2000)9999:9999<::AIDASI1039>3.0.CO;2-O
Wilder, A.J. Usability of Government Websites. A
Master’s Paper for the M.S. in L.S degree. April, 2007.
Leonidas Anthopoulos is an Associate Professor at the Business School of the TEI of Thessaly, Greece. Dr.
Anthopoulos has IT research, planning and Management experience with the development and deployment
of municipal and organizational IT environments. At his previous job positions, as an Expert Counselor at
the Hellenic Ministry of Foreign Affairs in e-Government and e-Diplomacy areas, as an IT researcher and
manager at the Research Committee of the Aristotle University of Thessaloniki (Greece), Municipality of
Trikala (Greece), Administration of Secondary Education of Trikala (Greece) and Information Society S.A.
(Greece), he was responsible for planning and managing the development of multiple IT systems for Greek
Government and for various Public Organizations. Among them it worth mentioning the Digital City of
Trikala (e-Trikala) project, the SYZEFXIS and Police-online projects, the central portal for the Hellenic
Ministry of Foreign Affairs etc. He is the author of several articles published on prestigious scientific journals, books and international conferences. His research interests concern, among others, e-Government,
Enterprise Architecture, Social Networks, etc.
Kleanthis Sirakoulis, Assistant Professor at Business School, TEI of Thessaly, Greece, School of Business
and Economics. Dr. Syrakoulis has extensive mathematics and project management experience mainly
focused on scheduling techniques, resource constraint management, statistics and non-governmental organizations (NGOs). He worked, as business manager and consultant for many public organizations. He
is the author of many articles published on prestigious scientific journals. His research interests include:
Project Management, Scheduling, NGOs, etc.
Copyright © 2015, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited.