32 1 NLOC_DMG,VAR_REG, THK, NEL,
33 2 OFF, AREA, NC1, NC2,
34 3 NC3, NC4, BUFNL, IMAT,
35 4 NDDL, ITASK, NG, JFT,
37 6 Y24, DT2T, THK0, AREA0,
47#include "implicit_f.inc"
59 INTEGER,
INTENT(IN) :: NFT
60 INTEGER :: NEL,IMAT,NDDL,ITASK,NG,
62 INTEGER,
DIMENSION(NEL) :: NC1,NC2,NC3,NC4
63 my_real,
DIMENSION(NEL,NDDL),
INTENT(INOUT)::
65 my_real,
DIMENSION(NEL),
INTENT(IN) ::
66 .
area,off,thk,x13,y13,x24,y24,thk0,area0
68 TYPE(BUF_NLOC_) ,
TARGET :: BUFNL
69 my_real,
INTENT(INOUT) ::
74 INTEGER I,II,K,N1,N2,N3,N4,L_NLOC,J,NDNOD
77 . b1,b2,b3,b4,sspnl,le_max,maxstif,dtnod,dt2p,
78 . nth1, nth2, bth1, bth2, k1, k2, k12, ntn_unl1,
79 . ntn_unl2,ntn_unl3,ntn_unl4,ntn_vnl1,ntn_vnl2,ntn_vnl3,
80 . ntn_vnl4,a1,a2,a3,a4
82 . vpg(2,4),pg1,pg,ksi,eta,sf1,sf2,sf3,sf4
83 parameter(pg=.577350269189626)
84 parameter(pg1=-.577350269189626)
85 my_real,
DIMENSION(:,:),
ALLOCATABLE ::
86 . f1,f2,f3,f4,sti1,sti2,sti3,sti4
87 my_real,
DIMENSION(:),
ALLOCATABLE ::
88 . vol,btb11,btb12,btb22
89 INTEGER,
DIMENSION(:),
ALLOCATABLE ::
91 my_real,
POINTER,
DIMENSION(:) ::
92 . vnl,fnl,unl,stifnl,mass,mass0,vnl0
93 my_real,
POINTER,
DIMENSION(:,:) ::
94 . massth,unlth,vnlth,fnlth
95 my_real,
DIMENSION(:,:),
ALLOCATABLE ::
97 DATA vpg/pg1,pg1,pg,pg1,pg,pg,pg1,pg/
103 l2 = nloc_dmg%LEN(imat)**2
104 xi = nloc_dmg%DAMP(imat)
105 zeta = nloc_dmg%DENS(imat)
106 sspnl = nloc_dmg%SSPNL(imat)
107 l_nloc = nloc_dmg%L_NLOC
108 le_max = nloc_dmg%LE_MAX(imat)
110 ALLOCATE(f1(nel,nddl),f2(nel,nddl),f3(nel,nddl),f4(nel,nddl))
114 ALLOCATE(sti1(nel,nddl),sti2(nel,nddl),sti3(nel,nddl),sti4(nel,nddl))
116 mass => nloc_dmg%MASS(1:l_nloc)
118 mass0 => nloc_dmg%MASS0(1:l_nloc)
120 ALLOCATE(vol(nel),btb11(nel),btb12(nel),btb22(nel),
121 . pos1(nel),pos2(nel),pos3(nel),pos4(nel))
123 vnl => nloc_dmg%VNL(1:l_nloc)
124 vnl0 => nloc_dmg%VNL_OLD(1:l_nloc)
125 unl => nloc_dmg%UNL(1:l_nloc)
133# include "vectorize.inc"
137 n1 = nloc_dmg%IDXI(nc1(i))
138 n2 = nloc_dmg%IDXI(nc2(i))
139 n3 = nloc_dmg%IDXI(nc3(i))
140 n4 = nloc_dmg%IDXI(nc4(i))
143 pos1(i) = nloc_dmg%POSI(n1)
144 pos2(i) = nloc_dmg%POSI(n2)
145 pos3(i) = nloc_dmg%POSI(n3)
146 pos4(i) = nloc_dmg%POSI(n4)
149 vol(i) = fourth*thk(i)*
area(i)
152 btb11(i) = y24(i)**2 + (-x24(i))**2
153 btb12(i) = y24(i)*(-y13(i)) + (-x24(i))*x13(i)
154 btb22(i) = (-y13(i))**2 + x13(i)**2
162 IF ((nddl > 1).AND.(l2>zero))
THEN
167 ALLOCATE(stifnlth(nel,nddl+1))
168 ALLOCATE(dtn(nel,nddl+1))
173 ALLOCATE(stifnlth(nel,nddl))
174 ALLOCATE(dtn(nel,nddl))
180 massth => bufnl%MASSTH(1:nel,1:ndnod)
181 unlth => bufnl%UNLTH(1:nel,1:ndnod)
182 vnlth => bufnl%VNLTH(1:nel,1:ndnod)
183 fnlth => bufnl%FNLTH(1:nel,1:ndnod)
200 IF ((nddl==2).AND.(k==2))
THEN
201 nth1 = (z0(k,nddl) - zth(k,nddl)) / (zth(k-1,nddl) - zth(k,nddl))
202 nth2 = (z0(k,nddl) - zth(k-1,nddl)) / (zth(k,nddl) - zth(k-1,nddl))
204 nth1 = (z0(k,nddl) - zth(k+1,nddl)) / (zth(k,nddl) - zth(k+1,nddl))
205 nth2 = (z0(k,nddl) - zth(k,nddl)) / (zth(k+1,nddl) - zth(k,nddl))
211 IF ((nddl==2).AND.(k==2))
THEN
212 bth1 = (one/(zth(k-1,nddl) - zth(k,nddl)))*(one/thk(i))
213 bth2 = (one/(zth(k,nddl) - zth(k-1,nddl)))*(one/thk(i))
215 bth1 = (one/(zth(k,nddl) - zth(k+1,nddl)))*(one/thk(i))
216 bth2 = (one/(zth(k+1,nddl) - zth(k,nddl
220 k1 = l2*(bth1**2) + nth1**2
221 k12 = l2*(bth1*bth2)+ (nth1*nth2)
222 k2 = l2*(bth2**2) + nth2**2
225 IF ((nddl==2).AND.(k==2))
THEN
226 fnlth(i,k-1) = fnlth(i,k-1) + (k1*unlth(i,k-1) + k12*unlth(i,k)
227 . + xi*((nth1**2)*vnlth(i,k-1)
228 . + (nth1*nth2)*vnlth(i,k))
229 . - (nth1*var_reg(i,k)))*vol(i)*wf(k,nddl)
230 fnlth(i,k) = fnlth(i,k) + (k12*unlth(i,k-1) + k2*unlth(i,k)
231 . + xi*(nth1*nth2*vnlth(i,k-1)
232 . + (nth2**2)*vnlth(i,k))
233 . - (nth2*var_reg(i,k)))*vol(i)*wf(k,nddl)
235 fnlth(i,k) = fnlth(i,k) + (k1*unlth(i,k) + k12*unlth(i,k+1)
236 . + xi*((nth1**2)*vnlth(i,k)
237 . + (nth1*nth2)*vnlth(i,k+1))
238 . - (nth1*var_reg(i,k)))*vol(i)*wf(k,nddl)
239 fnlth(i,k+1) = fnlth(i,k+1) + (k12*unlth(i,k) + k2*unlth(i,k+1)
240 . + xi*((nth1*nth2)*vnlth(i,k)
241 . + (nth2**2)*vnlth(i,k+1))
242 . - (nth2*var_reg(i,k)))*vol(i)*wf(k,nddl)
247 IF ((nddl==2).AND.(k==2))
THEN
248 stifnlth(i,k-1) = stifnlth(i,k-1) + (
max(abs(k1)+abs(k12),abs(k12)+abs(k2)))*vol(i)*wf(k,nddl)
249 stifnlth(i,k) = stifnlth(i,k) + (
max(abs(k1)+abs(k12),abs(k12)+abs(k2)))*vol(i)*wf(k,nddl)
251 stifnlth(i,k) = stifnlth(i,k) + (
max(abs(k1)+abs(k12),abs(k12)+abs(k2)))*vol(i)*wf(k,nddl)
252 stifnlth(i,k+1) = stifnlth(i,k+1) + (
max(abs(k1)+abs(k12),abs(k12)+abs(k2)))*vol(i)*wf(k,nddl)
266 dtn(i,k) = dtfac1(11)*cdamp*sqrt(two*massth(i,k)/
max(stifnlth(i,k),em20))
267 dtnod =
min(dtn(i,k),dtnod)
272 IF ((idtmin(11)==3).OR.(idtmin(11)==4).OR.(idtmin(11)==8))
THEN
274 IF (dtnod < dtmin1(11)*(sqrt(csta)))
THEN
277 IF (dtn(i,k) < dtmin1(11))
THEN
278 dt2p = dtmin1(11)/(dtfac1(11)*cdamp)
279 massth(i,k) =
max(massth(i,k),csta*half*stifnlth(i,k)*dt2p*dt2p*onep00001)
283 dtnod = dtmin1(11)*(sqrt(csta))
288 IF (dtnod < dt2t)
THEN
289 dt2t =
min(dt2t,dtnod)
296 vnlth(i,k) = vnlth(i,k) - (fnlth(i,k)/massth(i,k))*dt12
303 unlth(i,k) = unlth(i,k) + vnlth(i,k)*dt1
310 IF ((nddl==2).AND.(k==2))
THEN
311 nth1 = (z0(k,nddl) - zth(k,nddl))/(zth(k-1,nddl) - zth(k,nddl))
312 nth2 = (z0(k,nddl) - zth(k-1,nddl)) /(zth(k,nddl) - zth(k-1,nddl))
314 nth1 = (z0(k,nddl) - zth(k+1,nddl))/(zth(k,nddl) - zth(k+1,nddl))
315 nth2 = (z0(k,nddl) - zth(k,nddl)) /(zth(k+1,nddl) - zth(k,nddl))
320 IF ((nddl==2).AND.(k==2))
THEN
321 var_reg(i,k) = nth1*unlth(i,k-1) + nth2*unlth(i,k)
323 var_reg(i,k) = nth1*unlth(i,k) + nth2*unlth(i,k+1)
336 sf1 = fourth*(1-ksi)*(1-eta)
337 sf2 = fourth*(1+ksi)*(1-eta)
338 sf3 = fourth*(1+ksi)*(1+eta)
339 sf4 = fourth*(1-ksi)*(1+eta)
345# include "vectorize.inc"
349 IF (off(i) /= zero)
THEN
353 b1 = l2*vol(i)*wf(k,nddl)*(btb11(i)*unl(pos1(i)+k-1) + btb12(i)*unl(pos2(i)+k-1)
354 . - btb11(i)*unl(pos3(i)+k-1) - btb12(i)*unl(pos4(i)+k-1))
356 b2 = l2*vol(i)*wf(k,nddl)*(btb12(i)*unl(pos1(i)+k-1) + btb22(i)*unl(pos2(i)+k-1)
357 . - btb12(i)*unl(pos3(i)+k-1) - btb22(i)*unl(pos4(i)+k-1))
359 b3 = l2*vol(i)*wf(k,nddl)*(-btb11(i)*unl(pos1(i)+k-1) - btb12(i)*unl(pos2(i)+k-1)
360 . + btb11(i)*unl(pos3(i)+k-1) +
362 b4 = l2*vol(i)*wf(k,nddl)*(-btb12(i)*unl(pos1(i)+k-1) - btb22(i)*unl(pos2(i)+k-1)
363 . + btb12(i)*unl(pos3(i)+k-1) + btb22(i)*unl(pos4(i)+k-1))
366 ntn_unl1 = sf1*sf1*unl(pos1(i)+k-1) + sf1*sf2*unl(pos2(i)+k-1) + sf1*sf3*unl(pos3(i)+k-1
367 ntn_unl2 = sf2*sf1*unl(pos1(i)+k-1) + sf2*sf2*unl(pos2(i)+k-1)
368 ntn_unl3 = sf3*sf1*unl(pos1(i)+k-1) + sf3*sf2*unl(pos2(i)+k-1) + sf3*sf3*unl(pos3(i)+k-1) + sf3*sf4*unl(pos4(i)+k-1)
369 ntn_unl4 = sf4*sf1*unl(pos1(i)+k-1) + sf4*sf2*unl(pos2(i)+k-1) + sf4*sf3*unl(pos3(i)+k-1) + sf4*sf4*unl(pos4(i)+k-1)
372 ntn_vnl1 = sf1*sf1*vnl(pos1(i)+k-1) + sf1*sf2*vnl(pos2(i)+k-1) + sf1*sf3*vnl(pos3(i)+k-1) + sf1*sf4*vnl(pos4(i)+k-1)
374 ntn_vnl1 =
min(sqrt(mass(pos1(i)+k-1)/mass0(pos1(i)+k-1)),
375 . sqrt(mass(pos2(i)+k-1)/mass0(pos2(i)+k-1)),
376 . sqrt(mass(pos3(i)+k-1)/mass0(pos3(i)+k-1)),
377 . sqrt(mass(pos4(i)+k-1)/mass0(pos4(i)+k-1)))*ntn_vnl1
379 ntn_vnl2 = sf2*sf1*vnl(pos1(i)+k-1) + sf2*sf2*vnl(pos2(i)+k-1) + sf2*sf3*vnl(pos3(i)+k-1) + sf2*sf4*vnl(pos4(i)+k-1)
381 ntn_vnl2 =
min(sqrt(mass(pos1(i)+k-1)/mass0(pos1(i)+k-1)),
382 . sqrt(mass(pos2(i)+k-1)/mass0(pos2(i)+k-1)),
383 . sqrt(mass(pos3(i)+k-1)/mass0(pos3(i)+k-1)),
384 . sqrt(mass(pos4(i)+k-1)/mass0(pos4(i)+k-1)))*ntn_vnl2
386 ntn_vnl3 = sf3*sf1*vnl(pos1(i)+k-1) + sf3*sf2*vnl(pos2(i)+k-1) + sf3*sf3*vnl(pos3(i)+k-1) + sf3*sf4*vnl
388 ntn_vnl3 =
min(sqrt(mass(pos1(i)+k-1)/mass0(pos1(i)+k-1)),
389 . sqrt(mass(pos2(i)+k-1)/mass0(pos2(i)+k-1)),
390 . sqrt(mass(pos3(i)+k-1)/mass0(pos3(i)+k-1)),
391 . sqrt(mass(pos4(i)+k-1)/mass0(pos4(i)+k-1)))*ntn_vnl3
393 ntn_vnl4 = sf4*sf1*vnl(pos1(i)+k-1) + sf4*sf2*vnl(pos2(i)+k-1) + sf4*sf3*vnl
395 ntn_vnl4 =
min(sqrt(mass(pos1(i)+k-1)/mass0(pos1(i)+k-1)),
396 . sqrt(mass(pos2(i)+k-1)/mass0(pos2(i)+k-1)),
397 . sqrt(mass(pos3(i)+k-1)/mass0(pos3(i)+k-1)),
398 . sqrt(mass(pos4(i)+k-1)/mass0(pos4(i)+k-1)))*ntn_vnl4
402 a1 = (ntn_unl1 + ntn_vnl1*xi - sf1*var_reg(i,k)) * vol(i) * wf(k,nddl)
403 a2 = (ntn_unl2 + ntn_vnl2*xi - sf2*var_reg(i,k)) * vol(i) * wf(k,nddl)
404 a3 = (ntn_unl3 + ntn_vnl3*xi - sf3*var_reg(i,k)) * vol(i) * wf(k,nddl)
405 a4 = (ntn_unl4 + ntn_vnl4*xi - sf4*var_reg(i,k)) * vol(i) * wf(k,nddl)
415 sti1(i,k) = (abs(l2*btb11(i) + sf1*sf1) + abs(l2*btb12(i) + sf1*sf2) +
416 . abs(-l2*btb11(i) + sf1*sf3) + abs(-l2*btb12(i) + sf1*sf4))*vol(i)*wf(k,nddl)
417 sti2(i,k) = (abs(l2*btb12(i) + sf2*sf1) + abs(l2*btb22(i) + sf2*sf2) +
418 . abs(-l2*btb12(i) + sf2*sf3) + abs(-l2*btb22(i) + sf2*sf4))*vol(i)*wf(k,nddl)
419 sti3(i,k) = (abs(-l2*btb11(i) + sf3*sf1) + abs(-l2*btb12(i) + sf3*sf2) +
420 . abs(l2*btb11(i) + sf3*sf3) + abs(l2*btb12(i) + sf3*sf4))*vol(i)*wf(k,nddl)
421 sti4(i,k) = (abs(-l2*btb12(i) + sf4*sf1) + abs(-l2*btb22(i) + sf4*sf2) +
422 . abs(l2*btb12(i) + sf4*sf3) + abs(l2*btb22(i) + sf4*sf4))*vol(i)*wf(k
429 f1(i,k) = sqrt(mass(pos1(i)+k-1)/mass0(pos1(i)+k-1))*sf1*wf(k,nddl)*zeta*sspnl
430 . half*(vnl(pos1(i)+k-1)+vnl0(pos1(i)+k-1))*sqrt(area0(i))*thk0(i)
431 f2(i,k) = sqrt(mass(pos2(i)+k-1)/mass0(pos2(i)+k-1))*sf2*wf(k,nddl)*zeta*sspnl*
432 . half*(vnl(pos2(i)+k-1)+vnl0(pos2(i)+k-1))*sqrt(area0(i))*thk0(i)
433 f3(i,k) = sqrt(mass(pos3(i)+k-1)/mass0(pos3(i)+k-1))*sf3*wf(k,nddl)*zeta*sspnl*
434 . half*(vnl(pos3(i)+k-1)+vnl0(pos3(i)+k-1))*sqrt(area0
435 f4(i,k) = sqrt(mass(pos4(i)+k-1)/mass0(pos4(i)+k-1))*sf4*wf(k,nddl)*zeta*sspnl*
436 . half*(vnl(pos4(i)+k-1)+vnl0(pos4(i)+k-1))*sqrt(area0(i))*thk0(i
444 f1(i,k) = sf1*wf(k,nddl)*zeta*sspnl*half*(vnl(pos1(i)+k-1)+vnl0(pos1(i)+k-1))*sqrt(area0(i))*thk0(i)
445 f2(i,k) = sf2*wf(k,nddl)*zeta*sspnl*half*(vnl(pos2(i)+k-1)+vnl0(pos2(i)+k-1))*sqrt(area0(i))*thk0(i)
446 f3(i,k) = sf3*wf(k,nddl)*zeta*sspnl*half*(vnl(pos3(i)+k-1)+vnl0(pos3(i)+k-1))*sqrt(area0(i))*thk0
447 f4(i,k) = sf4*wf(k,nddl)*zeta*sspnl*half*(vnl(pos4(i)+k-1)+vnl0(pos4(i)+k-1))*sqrt(area0(i))*thk0(i)
458 IF (iparit == 0)
THEN
460 fnl => nloc_dmg%FNL(1:l_nloc,itask+1)
461 IF (nodadt > 0) stifnl => nloc_dmg%STIFNL(1:l_nloc,itask+1)
465# include "vectorize.inc"
468 fnl(pos1(i)+k-1) = fnl(pos1(i)+k-1) - f1(i,k)
469 fnl(pos2(i)+k-1) = fnl(pos2(i)+k-1) - f2(i,k)
470 fnl(pos3(i)+k-1) = fnl(pos3(i)+k-1) - f3(i,k)
471 fnl(pos4(i)+k-1) = fnl(pos4(i)+k-1) - f4(i,k)
474 maxstif =
max(sti1(i,k),sti2(i,k),sti3(i,k),sti4(i,k))
476 stifnl(pos1(i)+k-1) = stifnl(pos1(i)+k-1) + maxstif
477 stifnl(pos2(i)+k-1) = stifnl(pos2(i)+k-1) + maxstif
478 stifnl(pos3(i)+k-1) = stifnl(pos3(i)+k-1) + maxstif
479 stifnl(pos4(i)+k-1) = stifnl(pos4(i)+k-1) + maxstif
496 maxstif =
max(sti1(i,j),sti2(i,j),sti3(i,j),sti4(i,j))
499 k = nloc_dmg%IADC(1,ii)
501 nloc_dmg%FSKY(k,j) = -f1(i,j)
502 IF (nodadt > 0) nloc_dmg%STSKY(k,j) = maxstif
504 nloc_dmg%FSKY(k,j) = nloc_dmg%FSKY(k,j) - f1(i,j)
505 IF (nodadt > 0) nloc_dmg%STSKY(k,j) = nloc_dmg%STSKY(k,j) + maxstif
508 k = nloc_dmg%IADC(2,ii)
510 nloc_dmg%FSKY(k,j) = -f2(i,j)
511 IF (nodadt > 0) nloc_dmg%STSKY(k,j) = maxstif
513 nloc_dmg%FSKY(k,j) = nloc_dmg%FSKY(k,j) - f2(i,j)
514 IF (nodadt > 0) nloc_dmg%STSKY(k,j) = nloc_dmg%STSKY(k,j) + maxstif
517 k = nloc_dmg%IADC(3,ii)
519 nloc_dmg%FSKY(k,j) = -f3(i,j)
520 IF (nodadt > 0) nloc_dmg%STSKY(k,j) = maxstif
522 nloc_dmg%FSKY(k,j) = nloc_dmg%FSKY(k,j) - f3(i,j)
523 IF (nodadt > 0) nloc_dmg%STSKY(k,j) = nloc_dmg%STSKY(k,j) + maxstif
526 k = nloc_dmg%IADC(4,ii)
528 nloc_dmg%FSKY(k,j) = -f4(i,j)
529 IF (nodadt > 0) nloc_dmg%STSKY(k,j) = maxstif
531 nloc_dmg%FSKY(k,j) = nloc_dmg%FSKY(k,j) - f4(i,j)
532 IF (nodadt > 0) nloc_dmg%STSKY(k,j) = nloc_dmg%STSKY(k,j) + maxstif
541 IF (
ALLOCATED(f1))
DEALLOCATE(f1)
542 IF (
ALLOCATED(f2))
DEALLOCATE(f2)
543 IF (
ALLOCATED(f3))
DEALLOCATE(f3)
544 IF (
ALLOCATED(f4))
DEALLOCATE(f4)
545 IF (
ALLOCATED(stifnlth))
DEALLOCATE(stifnlth)
546 IF (
ALLOCATED(btb11))
DEALLOCATE(btb11)
547 IF (
ALLOCATED(btb12))
DEALLOCATE(btb12)
548 IF (
ALLOCATED(btb22))
DEALLOCATE(btb22)
549 IF (
ALLOCATED(pos1))
DEALLOCATE(pos1)
550 IF (
ALLOCATED(pos2))
DEALLOCATE(pos2)
551 IF (
ALLOCATED(pos3))
DEALLOCATE(pos3)
552 IF (
ALLOCATED(pos4))
DEALLOCATE(pos4)
553 IF (
ALLOCATED(sti1))
DEALLOCATE(sti1)
554 IF (
ALLOCATED(sti2))
DEALLOCATE(sti2)
555 IF (
ALLOCATED(sti3))
DEALLOCATE(sti3)
556 IF (
ALLOCATED(sti4))
DEALLOCATE(sti4)