Skip to content

Commit e66dd68

Browse files
Merge pull request #491 from de-souza/fix_debugging
Fix debugging docs
2 parents a5bfc91 + aba3c00 commit e66dd68

File tree

1 file changed

+12
-10
lines changed

1 file changed

+12
-10
lines changed

docs/src/pinn/debugging.md

Lines changed: 12 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -50,15 +50,17 @@ isapprox(dphi[1][2], dphi2, atol=1e-8)
5050

5151
indvars = [x,t]
5252
depvars = [u(x, t)]
53+
dict_depvars_input = Dict(:u => [:x, :t])
5354
dim = length(domains)
5455
dx = 0.1
5556
strategy = NeuralPDE.GridTraining(dx)
57+
integral = NeuralPDE.get_numeric_integral(strategy, indvars, depvars, chain, derivative)
5658

57-
_pde_loss_function = NeuralPDE.build_loss_function(eq,indvars,depvars,phi,derivative,chain,initθ,strategy)
59+
_pde_loss_function = NeuralPDE.build_loss_function(eq,indvars,depvars,phi,derivative,integral,chain,initθ,strategy)
5860

59-
julia> expr_pde_loss_function = NeuralPDE.build_symbolic_loss_function(eq,indvars,depvars,phi,derivative,chain,initθ,strategy)
61+
julia> expr_pde_loss_function = NeuralPDE.build_symbolic_loss_function(eq,indvars,depvars,dict_depvars_input,phi,derivative,integral,chain,initθ,strategy)
6062

61-
:((cord, var"##θ#529", phi, derivative, u)->begin
63+
:((cord, var"##θ#529", phi, derivative, integral, u)->begin
6264
begin
6365
let (x, t) = (cord[[1], :], cord[[2], :])
6466
derivative.(phi, u, cord, Array{Float32,1}[[0.0, 0.0049215667], [0.0, 0.0049215667]], 2, var"##θ#529") .- derivative.(phi, u, cord, Array{Float32,1}[[0.0049215667, 0.0], [0.0049215667, 0.0]], 2, var"##θ#529")
@@ -74,35 +76,35 @@ julia> bc_indvars = NeuralPDE.get_variables(bcs,indvars,depvars)
7476
[:x]
7577

7678
_bc_loss_functions = [NeuralPDE.build_loss_function(bc,indvars,depvars,
77-
phi,derivative,chain,initθ,strategy,
79+
phi,derivative,integral,chain,initθ,strategy,
7880
bc_indvars = bc_indvar) for (bc,bc_indvar) in zip(bcs,bc_indvars)]
7981

80-
julia> expr_bc_loss_functions = [NeuralPDE.build_symbolic_loss_function(bc,indvars,depvars,
81-
phi,derivative,chain,initθ,strategy,
82+
julia> expr_bc_loss_functions = [NeuralPDE.build_symbolic_loss_function(bc,indvars,depvars,dict_depvars_input,
83+
phi,derivative,integral,chain,initθ,strategy,
8284
bc_indvars = bc_indvar) for (bc,bc_indvar) in zip(bcs,bc_indvars)]
8385
4-element Array{Expr,1}:
84-
:((cord, var"##θ#529", phi, derivative, u)->begin
86+
:((cord, var"##θ#529", phi, derivative, integral, u)->begin
8587
begin
8688
let (x, t) = (cord[[1], :], cord[[2], :])
8789
u.(cord, var"##θ#529", phi) .- 0.0
8890
end
8991
end
9092
end)
91-
:((cord, var"##θ#529", phi, derivative, u)->begin
93+
:((cord, var"##θ#529", phi, derivative, integral, u)->begin
9294
begin
9395
let (x, t) = (cord[[1], :], cord[[2], :])
9496
u.(cord, var"##θ#529", phi) .- 0.0
9597
end
9698
end
9799
end)
98-
:((cord, var"##θ#529", phi, derivative, u)->begin
100+
:((cord, var"##θ#529", phi, derivative, integral, u)->begin
99101
begin
100102
let (x, t) = (cord[[1], :], cord[[2], :])
101103
u.(cord, var"##θ#529", phi) .- (*).(x, (+).(1.0, (*).(-1, x)))
102104
end
103105
end
104106
end)
105-
:((cord, var"##θ#529", phi, derivative, u)->begin
107+
:((cord, var"##θ#529", phi, derivative, integral, u)->begin
106108
begin
107109
let (x, t) = (cord[[1], :], cord[[2], :])
108110
derivative.(phi, u, cord, Array{Float32,1}[[0.0, 0.0049215667]], 1, var"##θ#529") .- 0.0

0 commit comments

Comments
 (0)