Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
P
python3-imagej-tiff
Project
Project
Details
Activity
Releases
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Elphel
python3-imagej-tiff
Commits
2d9f3013
Commit
2d9f3013
authored
Aug 28, 2018
by
Oleg Dzhimiev
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
fixed shaping error
parent
758c8a5d
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
7 additions
and
7 deletions
+7
-7
nn_ds_neibs1_tmp.py
nn_ds_neibs1_tmp.py
+7
-7
No files found.
nn_ds_neibs1_tmp.py
View file @
2d9f3013
...
@@ -55,7 +55,7 @@ ONLY_TILE = None # 4 # None # 0 # 4# None # (remove all but center tile
...
@@ -55,7 +55,7 @@ ONLY_TILE = None # 4 # None # 0 # 4# None # (remove all but center tile
ZIP_LHVAR
=
True
# combine _lvar and _hvar as odd/even elements
ZIP_LHVAR
=
True
# combine _lvar and _hvar as odd/even elements
#DEBUG_PACK_TILES = True
#DEBUG_PACK_TILES = True
WLOSS_LAMBDA
=
0.
00
1
# 5.0 # 1.0 # fraction of the W_loss (input layers weight non-uniformity) added to G_loss
WLOSS_LAMBDA
=
0.1
# 5.0 # 1.0 # fraction of the W_loss (input layers weight non-uniformity) added to G_loss
SUFFIX
=
str
(
NET_ARCH1
)
+
'-'
+
str
(
NET_ARCH2
)
+
([
"R"
,
"A"
][
ABSOLUTE_DISPARITY
])
SUFFIX
=
str
(
NET_ARCH1
)
+
'-'
+
str
(
NET_ARCH2
)
+
([
"R"
,
"A"
][
ABSOLUTE_DISPARITY
])
# CLUSTER_RADIUS should match input data
# CLUSTER_RADIUS should match input data
CLUSTER_RADIUS
=
1
# 1 - 3x3, 2 - 5x5 tiles
CLUSTER_RADIUS
=
1
# 1 - 3x3, 2 - 5x5 tiles
...
@@ -476,7 +476,7 @@ def network_summary_w_b(scope, in_shape, out_shape, layout, index, network_scope
...
@@ -476,7 +476,7 @@ def network_summary_w_b(scope, in_shape, out_shape, layout, index, network_scope
wt
=
tf
.
transpose
(
w
,[
1
,
0
])
wt
=
tf
.
transpose
(
w
,[
1
,
0
])
wt
=
wt
[:,:
-
1
]
wt
=
wt
[:,:
-
1
]
tmp1
=
[]
tmp1
=
[]
for
i
in
range
(
layout
[
index
]
):
for
i
in
range
(
out_shape
):
# reset when even
# reset when even
if
i
%
2
==
0
:
if
i
%
2
==
0
:
...
@@ -511,7 +511,7 @@ def network_summary_w_b(scope, in_shape, out_shape, layout, index, network_scope
...
@@ -511,7 +511,7 @@ def network_summary_w_b(scope, in_shape, out_shape, layout, index, network_scope
tmp1
.
append
(
ts
)
tmp1
.
append
(
ts
)
imsum1
=
tf
.
concat
(
tmp1
,
axis
=
0
)
imsum1
=
tf
.
concat
(
tmp1
,
axis
=
0
)
imsum1_1
=
tf
.
reshape
(
imsum1
,[
1
,
layout
[
index
]
*
(
TILE_SIDE
+
1
)
//
2
,
2
*
TILE_LAYERS
*
(
TILE_SIDE
+
1
),
3
])
imsum1_1
=
tf
.
reshape
(
imsum1
,[
1
,
out_shape
*
(
TILE_SIDE
+
1
)
//
2
,
2
*
TILE_LAYERS
*
(
TILE_SIDE
+
1
),
3
])
tf
.
summary
.
image
(
"sub_w8s"
,
imsum1_1
)
tf
.
summary
.
image
(
"sub_w8s"
,
imsum1_1
)
...
@@ -547,7 +547,7 @@ def network_summary_w_b(scope, in_shape, out_shape, layout, index, network_scope
...
@@ -547,7 +547,7 @@ def network_summary_w_b(scope, in_shape, out_shape, layout, index, network_scope
missing_in_block
=
math
.
pow
(
block_side
,
2
)
-
block_size
missing_in_block
=
math
.
pow
(
block_side
,
2
)
-
block_size
tmp1
=
[]
tmp1
=
[]
for
i
in
range
(
layout
[
index
]
):
for
i
in
range
(
out_shape
):
# reset when even
# reset when even
if
i
%
4
==
0
:
if
i
%
4
==
0
:
...
@@ -600,7 +600,7 @@ def network_summary_w_b(scope, in_shape, out_shape, layout, index, network_scope
...
@@ -600,7 +600,7 @@ def network_summary_w_b(scope, in_shape, out_shape, layout, index, network_scope
tmp1
.
append
(
ts
)
tmp1
.
append
(
ts
)
imsum2
=
tf
.
concat
(
tmp1
,
axis
=
0
)
imsum2
=
tf
.
concat
(
tmp1
,
axis
=
0
)
tf
.
summary
.
image
(
"inter_w8s"
,
tf
.
reshape
(
imsum2
,[
1
,
layout
[
index
]
*
cluster_side
*
(
block_side
+
1
)
//
4
,
4
*
cluster_side
*
(
block_side
+
1
),
3
]))
tf
.
summary
.
image
(
"inter_w8s"
,
tf
.
reshape
(
imsum2
,[
1
,
out_shape
*
cluster_side
*
(
block_side
+
1
)
//
4
,
4
*
cluster_side
*
(
block_side
+
1
),
3
]))
...
@@ -627,7 +627,7 @@ def network_sub(input, layout, reuse, sym8 = False):
...
@@ -627,7 +627,7 @@ def network_sub(input, layout, reuse, sym8 = False):
if
not
reuse_this
:
if
not
reuse_this
:
with
tf
.
variable_scope
(
scp
,
reuse
=
True
)
:
# tf.AUTO_REUSE):
with
tf
.
variable_scope
(
scp
,
reuse
=
True
)
:
# tf.AUTO_REUSE):
inp_weights
.
append
(
tf
.
get_variable
(
'weights'
))
# ,shape=[inp.shape[1],num_outs]))
inp_weights
.
append
(
tf
.
get_variable
(
'weights'
))
# ,shape=[inp.shape[1],num_outs]))
network_summary_w_b
(
scp
,
inp
.
shape
[
1
],
num_
outs
,
layout
,
i
,
'sub'
)
network_summary_w_b
(
scp
,
inp
.
shape
[
1
],
num_
sym8
,
layout
,
i
,
'sub'
)
if
num_non_sum
>
0
:
if
num_non_sum
>
0
:
reuse_this
=
reuse
reuse_this
=
reuse
scp
=
'g_fc_sub'
+
str
(
i
)
+
"r"
scp
=
'g_fc_sub'
+
str
(
i
)
+
"r"
...
@@ -635,7 +635,7 @@ def network_sub(input, layout, reuse, sym8 = False):
...
@@ -635,7 +635,7 @@ def network_sub(input, layout, reuse, sym8 = False):
if
not
reuse_this
:
if
not
reuse_this
:
with
tf
.
variable_scope
(
scp
,
reuse
=
True
)
:
# tf.AUTO_REUSE):
with
tf
.
variable_scope
(
scp
,
reuse
=
True
)
:
# tf.AUTO_REUSE):
inp_weights
.
append
(
tf
.
get_variable
(
'weights'
))
# ,shape=[inp.shape[1],num_outs]))
inp_weights
.
append
(
tf
.
get_variable
(
'weights'
))
# ,shape=[inp.shape[1],num_outs]))
network_summary_w_b
(
scp
,
inp
.
shape
[
1
],
num_
outs
,
layout
,
i
,
'sub'
)
network_summary_w_b
(
scp
,
inp
.
shape
[
1
],
num_
non_sum
,
layout
,
i
,
'sub'
)
fc
.
append
(
tf
.
concat
(
fc_sym
,
1
,
name
=
'sym_input_layer'
))
fc
.
append
(
tf
.
concat
(
fc_sym
,
1
,
name
=
'sym_input_layer'
))
else
:
else
:
scp
=
'g_fc_sub'
+
str
(
i
)
scp
=
'g_fc_sub'
+
str
(
i
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment