Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
D
dune-istl
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Wiki
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Build
Pipelines
Jobs
Pipeline schedules
Artifacts
Deploy
Releases
Model registry
Operate
Environments
Monitor
Incidents
Service Desk
Analyze
Value stream analytics
Contributor analytics
CI/CD analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
Core Modules
dune-istl
Commits
9d95d20f
Commit
9d95d20f
authored
20 years ago
by
Markus Blatt
Browse files
Options
Downloads
Patches
Plain Diff
Building the remote index list in parallel should work now.
[[Imported from SVN: r101]]
parent
e9eee3d0
No related branches found
Branches containing commit
No related tags found
Tags containing commit
No related merge requests found
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
istl/indexset.hh
+22
-3
22 additions, 3 deletions
istl/indexset.hh
istl/remoteindices.hh
+25
-19
25 additions, 19 deletions
istl/remoteindices.hh
istl/test/indicestest.cc
+14
-12
14 additions, 12 deletions
istl/test/indicestest.cc
with
61 additions
and
34 deletions
istl/indexset.hh
+
22
−
3
View file @
9d95d20f
...
...
@@ -520,7 +520,7 @@ namespace Dune
template
<
class
TG
,
class
TL
,
int
N
>
IndexSet
<
TG
,
TL
,
N
>::
IndexSet
()
:
state_
(
GROUND
)
:
state_
(
GROUND
)
,
noPublic_
(
0
),
seqNo_
(
0
)
{}
template
<
class
TG
,
class
TL
,
int
N
>
...
...
@@ -594,6 +594,7 @@ namespace Dune
std
::
sort
(
newIndices_
.
begin
(),
newIndices_
.
end
());
merge
();
seqNo_
++
;
state_
=
GROUND
;
}
...
...
@@ -601,6 +602,13 @@ namespace Dune
inline
void
IndexSet
<
TG
,
TL
,
N
>::
merge
(){
if
(
localIndices_
.
size
()
==
0
)
{
iterator
added
=
newIndices_
.
begin
(),
endadded
=
newIndices_
.
end
();
while
(
added
!=
newIndices_
.
end
())
{
if
(
added
->
local
().
isPublic
())
noPublic_
++
;
++
added
;
}
localIndices_
=
newIndices_
;
newIndices_
.
clear
();
}
...
...
@@ -609,33 +617,44 @@ namespace Dune
ArrayList
<
IndexPair
<
TG
,
TL
>
,
N
>
tempPairs
;
iterator
old
=
localIndices_
.
begin
(),
endold
=
localIndices_
.
end
();
iterator
added
=
newIndices_
.
begin
(),
endadded
=
newIndices_
.
end
();
noPublic_
=
0
;
while
(
old
!=
localIndices_
.
end
()
&&
added
!=
newIndices_
.
end
())
{
if
(
old
->
local
().
state
()
==
DELETED
)
if
(
old
->
local
().
state
()
==
DELETED
)
{
old
.
eraseToHere
();
}
else
if
(
old
->
global
()
<
added
->
global
())
{
tempPairs
.
push_back
(
*
old
);
if
(
old
->
local
().
isPublic
())
noPublic_
++
;
old
.
eraseToHere
();
}
else
{
tempPairs
.
push_back
(
*
added
);
if
(
added
->
local
().
isPublic
())
noPublic_
++
;
added
.
eraseToHere
();
}
}
while
(
old
!=
localIndices_
.
end
())
{
if
(
old
->
local
().
state
()
!=
DELETED
)
if
(
old
->
local
().
state
()
!=
DELETED
)
{
if
(
old
->
local
().
isPublic
())
noPublic_
++
;
tempPairs
.
push_back
(
*
old
);
}
old
.
eraseToHere
();
}
while
(
added
!=
newIndices_
.
end
())
{
tempPairs
.
push_back
(
*
added
);
if
(
added
->
local
().
isPublic
())
noPublic_
++
;
added
.
eraseToHere
();
}
localIndices_
=
tempPairs
;
...
...
This diff is collapsed.
Click to expand it.
istl/remoteindices.hh
+
25
−
19
View file @
9d95d20f
...
...
@@ -425,14 +425,14 @@ namespace Dune {
if
(
type
==
MPI_DATATYPE_NULL
)
{
int
length
[
3
];
MPI_Aint
disp
[
3
];
MPI_Datatype
types
[
4
]
=
{
MPI_LB
,
MPITraits
<
char
>::
getType
(),
MPI_UB
};
MPI_Datatype
types
[
3
]
=
{
MPI_LB
,
MPITraits
<
char
>::
getType
(),
MPI_UB
};
ParallelLocalIndex
<
T
>
rep
[
2
];
length
[
0
]
=
length
[
1
]
=
length
[
2
]
=
1
;
MPI_Address
(
rep
,
disp
);
// lower bound of the datatype
MPI_Address
(
&
(
rep
[
0
].
attribute_
),
disp
+
1
);
MPI_Address
(
rep
+
1
,
disp
+
2
);
// upper bound od the datatype
for
(
int
i
=
1
;
i
<
3
;
i
++
)
disp
[
1
]
-=
disp
[
0
];
for
(
int
i
=
2
;
i
>=
0
;
--
i
)
disp
[
i
]
-=
disp
[
0
];
MPI_Type_struct
(
3
,
length
,
disp
,
types
,
&
type
);
MPI_Type_commit
(
&
type
);
}
...
...
@@ -456,8 +456,8 @@ namespace Dune {
MPI_Address
(
&
(
rep
[
0
].
global_
),
disp
+
1
);
MPI_Address
(
&
(
rep
[
0
].
local_
),
disp
+
2
);
MPI_Address
(
rep
+
1
,
disp
+
3
);
// upper bound of the datatype
for
(
int
i
=
1
;
i
<
4
;
i
++
)
disp
[
1
]
-=
disp
[
0
];
for
(
int
i
=
3
;
i
>=
0
;
--
i
)
disp
[
i
]
-=
disp
[
0
];
MPI_Type_struct
(
4
,
length
,
disp
,
types
,
&
type
);
MPI_Type_commit
(
&
type
);
}
...
...
@@ -481,8 +481,8 @@ namespace Dune {
MPI_Address
(
&
(
rep
[
0
].
globalIndex_
),
disp
+
1
);
MPI_Address
(
&
(
rep
[
0
].
attribute_
),
disp
+
2
);
MPI_Address
(
rep
+
1
,
disp
+
3
);
// upper bound od the datatype
for
(
int
i
=
1
;
i
<
4
;
i
++
)
disp
[
1
]
-=
disp
[
0
];
for
(
int
i
=
0
;
i
<
4
;
i
++
)
disp
[
i
]
-=
disp
[
0
];
MPI_Type_struct
(
4
,
length
,
disp
,
types
,
&
type
);
MPI_Type_commit
(
&
type
);
}
...
...
@@ -576,10 +576,13 @@ namespace Dune {
char
**
buffer
=
new
char
*
[
2
];
int
bufferSize
;
int
position
=
0
;
int
intSize
;
MPI_Pack_size
(
maxPublish
,
MPITraits
<
PairType
>::
getType
(),
comm_
,
&
bufferSize
);
MPI_Pack_size
(
1
,
MPI_INT
,
comm_
,
&
intSize
);
bufferSize
+=
intSize
;
buffer
[
0
]
=
new
char
[
bufferSize
];
buffer
[
1
]
=
new
char
[
bufferSize
];
...
...
@@ -596,11 +599,12 @@ namespace Dune {
//Now pack the indices
const
PairType
*
pair
=
myPairs
[
0
];
MPI_Datatype
type
=
MPITraits
<
PairType
>::
getType
();
for
(
const_iterator
index
=
source_
.
begin
();
index
!=
end
;
++
index
)
if
(
b
||
index
->
local
().
isPublic
())
{
MPI_Pack
(
const_cast
<
PairType
*>
(
&
(
*
index
)),
1
,
MPITraits
<
PairType
>::
getType
()
,
type
,
buffer
[
0
],
bufferSize
,
&
position
,
comm_
);
pair
=
&
(
*
index
);
++
pair
;
...
...
@@ -609,24 +613,25 @@ namespace Dune {
char
*
p_out
=
buffer
[
1
-
(
proc
%
2
)];
char
*
p_in
=
buffer
[
proc
%
2
];
int
receiveCount
=
bufferSize
;
MPI_Status
status
;
if
(
rank
%
2
==
0
)
{
MPI_Ssend
(
&
p_out
,
position
,
MPI_PACKED
,
(
rank
+
1
)
%
procs
,
MPI_Ssend
(
p_out
,
position
,
MPI_PACKED
,
(
rank
+
1
)
%
procs
,
commTag_
,
comm_
);
MPI_Recv
(
&
p_in
,
maxPublish
,
MPI_PACKED
,
(
rank
+
procs
-
1
)
%
procs
,
MPI_Recv
(
p_in
,
receiveCount
,
MPI_PACKED
,
(
rank
+
procs
-
1
)
%
procs
,
commTag_
,
comm_
,
&
status
);
}
else
{
MPI_Recv
(
&
p_in
,
maxPublish
,
MPI_PACKED
,
(
rank
+
procs
-
1
)
%
procs
,
MPI_Recv
(
p_in
,
receiveCount
,
MPI_PACKED
,
(
rank
+
procs
-
1
)
%
procs
,
commTag_
,
comm_
,
&
status
);
MPI_Send
(
&
p_out
,
position
,
MPI_PACKED
,
(
rank
+
1
)
%
procs
,
commTag_
,
comm_
);
MPI_S
s
end
(
p_out
,
position
,
MPI_PACKED
,
(
rank
+
1
)
%
procs
,
commTag_
,
comm_
);
}
// unpack the number of indices we
got
// unpack the number of indices we
received
int
noReceived
;
position
=
0
;
MPI_Unpack
(
&
p_in
,
maxPublish
,
&
position
,
&
noReceived
,
1
,
MPI_INT
,
comm_
);
MPI_Unpack
(
p_in
,
receiveCount
,
&
position
,
&
noReceived
,
1
,
MPI_INT
,
comm_
);
// The process these indices are from
int
remoteProc
=
(
rank
+
procs
-
proc
)
%
procs
;
SLList
<
RemoteIndex
<
GlobalIndexType
,
AttributeType
>
>&
remoteIndices
=
remoteIndices_
[
remoteProc
];
...
...
@@ -635,7 +640,7 @@ namespace Dune {
for
(
int
i
=
0
;
i
<
noReceived
;
i
++
)
{
RemoteIndexInformation
<
TG
,
TA
>
index
;
MPI_Unpack
(
&
p_in
,
maxPublish
,
&
position
,
&
index
,
1
,
MPI_Unpack
(
p_in
,
receiveCount
,
&
position
,
&
index
,
1
,
MPITraits
<
RemoteIndexInformation
<
TG
,
TA
>
>::
getType
(),
comm_
);
//Check if we know the global index
while
(
pairIndex
<
publish
)
{
...
...
@@ -657,7 +662,8 @@ namespace Dune {
template
<
class
TG
,
class
TA
>
inline
void
RemoteIndices
<
TG
,
TA
>::
rebuild
()
{
buildLocal
();
if
(
&
source_
!=
&
dest_
)
buildLocal
();
buildRemote
();
}
...
...
This diff is collapsed.
Click to expand it.
istl/test/indicestest.cc
+
14
−
12
View file @
9d95d20f
...
...
@@ -14,7 +14,7 @@ void testIndices()
using
namespace
Dune
;
// The global grid size
const
int
N
=
10
00
;
const
int
N
=
10
;
// Process configuration
int
procs
,
rank
;
MPI_Comm_size
(
MPI_COMM_WORLD
,
&
procs
);
...
...
@@ -27,7 +27,7 @@ void testIndices()
IndexSet
<
int
,
ParallelLocalIndex
<
GridFlags
>
>
distIndexSet
;
// global indexset
IndexSet
<
int
,
ParallelLocalIndex
<
GridFlags
>
>
&
globalIndexSet
=
distIndexSet
;
IndexSet
<
int
,
ParallelLocalIndex
<
GridFlags
>
>
*
globalIndexSet
;
// Set up the indexsets.
int
start
=
rank
*
n
;
...
...
@@ -37,7 +37,7 @@ void testIndices()
for
(
int
i
=
start
;
i
<
end
;
i
++
)
for
(
int
j
=
start
;
j
<
end
;
j
++
)
{
bool
isPublic
=
(
i
==
start
)
||
(
i
==
end
)
||
(
j
==
start
)
||
(
j
=
end
);
bool
isPublic
=
true
||
(
i
==
start
)
||
(
i
==
end
)
||
(
j
==
start
)
||
(
j
=
=
end
);
GridFlags
flag
=
owner
;
if
((
i
==
start
&&
i
!=
0
)
||
(
i
==
end
&&
i
!=
N
-
1
)
...
...
@@ -53,22 +53,24 @@ void testIndices()
// build global indexset on first process
if
(
rank
==
0
)
{
globalIndexSet
=
*
(
new
IndexSet
<
int
,
ParallelLocalIndex
<
GridFlags
>
>
()
)
;
globalIndexSet
.
beginResize
();
globalIndexSet
=
new
IndexSet
<
int
,
ParallelLocalIndex
<
GridFlags
>
>
();
globalIndexSet
->
beginResize
();
for
(
int
i
=
0
;
i
<
N
;
i
++
)
for
(
int
j
=
0
;
j
<
N
;
j
++
)
globalIndexSet
.
add
(
i
*
N
+
j
,
ParallelLocalIndex
<
GridFlags
>
(
owner
,
false
));
globalIndexSet
.
endResize
();
globalIndexSet
->
add
(
i
*
N
+
j
,
ParallelLocalIndex
<
GridFlags
>
(
owner
,
false
));
globalIndexSet
->
endResize
();
}
// Build remote indices
RemoteIndices
<
int
,
GridFlags
>
distRemote
(
distIndexSet
,
distIndexSet
,
MPI_COMM_WORLD
);
RemoteIndices
<
int
,
GridFlags
>
crossRemote
(
distIndexSet
,
globalIndexSet
,
MPI_COMM_WORLD
);
if
(
rank
==
0
)
{
RemoteIndices
<
int
,
GridFlags
>
crossRemote
(
distIndexSet
,
*
globalIndexSet
,
MPI_COMM_WORLD
);
delete
globalIndexSet
;
}
else
RemoteIndices
<
int
,
GridFlags
>
distRemote
(
distIndexSet
,
distIndexSet
,
MPI_COMM_WORLD
);
}
int
main
(
int
argc
,
char
**
argv
)
...
...
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment